<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>3D-CNN in Drug Resistance Detection and Tuberculosis Classi cation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jo~ao Figueira Silva</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jorge Miguel Silva</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Eduardo Pinho</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Carlos Costa</string-name>
          <email>carlos.costag@ua.pt</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>DETI - Institute of Electronics and Informatics Engineering of Aveiro University of Aveiro</institution>
          ,
          <country country="PT">Portugal</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Object classi cation is a very demanding eld in computer vision, especially when dealing with medical imaging datasets, which are often small and have unbalanced distributions. Deep learning (DL) methods have proven to be e ective in dealing with such problems and have established themselves as the state-of-the-art. ImageCLEFtuberculosis is a challenge that encompasses the classi cation problem on medical images, and is divided into two subtasks: Drug Resistance Detection and Tuberculosis classi cation. For both subtasks, provided images were pre-processed to segment the lungs from the CT volumes. Afterwards, pre-processed CT volumes were fed in batches to a 3D convolutional neural network. Test results for the Drug Resistance detection task scored an accuracy of 46.5% and AUC of 0.46, while in the Tuberculosis classication task an accuracy of 24% and Cohen's Kappa value of 0.022 were obtained. Using data augmentation and weight normalization, the overtting problem could be reduced, and submitted models' performance improved.</p>
      </abstract>
      <kwd-group>
        <kwd>3D-CNN</kwd>
        <kwd>Neural Networks</kwd>
        <kwd>Deep Learning</kwd>
        <kwd>Medical Imaging</kwd>
        <kwd>CT</kwd>
        <kwd>Tuberculosis</kwd>
        <kwd>ImageCLEF</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        The ImageCLEFtuberculosis task [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] from ImageCLEF 2017 [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] is a challenge
centered on medical imaging, that has the motivation of improving
tuberculosis treatment and reducing its impact on patients through the development of
systems capable of extracting the tuberculosis type and drug resistances from
image data alone. Usually, working with medical imaging datasets encompasses
distinct challenges such as the limited access to data, its reduced size and the
unbalanced distributions. Deep learning (DL) methods have been increasingly
explored in the eld of image analysis, with neural networks leading to major
breakthroughs in renown challenges, such as the MNIST Digit Image
Classication Problem and the ImageNet Large Scale Visual Recognition Challenge
(ILSVRC) [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], where they are considered the state-of-the-art [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. These networks
present great interest since they automatically learn high-level representations
from the data, and can be used to reduce the data dimensionality [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>
        In recent years, deep learning has started to make a signi cant appearance in
the eld of medical imaging with promising results [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Following this trend, this
article assesses the viability of this technology to solve ImageCLEF challenges
through the development of a 3D Convolutional Neural Network (CNN) model.
      </p>
      <p>The ImageCLEFtuberculosis task is divided into two separate and
independent subtasks: drug resistance detection and tuberculosis classi cation. The goal
of the rst task was to assess the probability of a tuberculosis (TB) patient
having a resistant form of tuberculosis based on the analysis of a chest CT scan,
whereas the second one focused on classifying the TB type from ve possible
types of TB.</p>
      <p>This article describes the proposed solution and runs submitted by the
Bioinformatics team for both subtasks. The developed methodology is presented in
Section 2, results are presented and discussed in Section 3, and nally Section 4
draws some conclusions and future work.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Methodology</title>
      <p>To address the MDR detection and TB type subtasks from
ImageCLEFtuberculosis, we propose a two-stage pipeline: Data pre-processing and a DL model
(Figure 1). The pre-processing stage was applied to both subtasks whereas the
DL model was ne-tuned for each subtask.</p>
      <p>CT Volumes</p>
      <p>Pre-Processing</p>
      <p>Deep Learning</p>
      <p>Output</p>
      <p>Pre-processing stage used the Computed Tomography (CT) images,
segmented the lungs, and resized data to be ready to feed the DL model. On the
other hand, the DL model used batches of pre-processed data and classi ed it.
Each of the stages is explained in more detail in the next subsections.</p>
      <p>An important aspect of proposed approach is related with the fact that a
CT volume is composed by several images and the observation that a single slice
might provide poor classi cation results. So, we decided to feed the DL models
with volumes composed of stacks of CT slices, option that conducted us to use
of a 3D-CNN model instead of a conventional CNN model. This option brought
also implications concerning the shape of the models input tensors, which were
solved in the data pre-processing step.</p>
      <sec id="sec-2-1">
        <title>Data Pre-processing</title>
        <p>
          Pre-processing stage has the responsibility of preparing data for posterior
processes, namely feeding the DL model. For the drug resistance detection subtask,
a train dataset with 230 CT volumes and a test dataset with 214 CT volumes
were provided. In this subtask data had two possible classes. Regarding the
tuberculosis classi cation subtask, a train dataset with 500 CT volumes and a
test dataset with 300 CT volumes were provided, with data having ve possible
classes. CT volumes had a variable number of slices, with slice size being 512x512
pixels [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ].
        </p>
        <p>In the training datasets, CT volumes had the lungs segmented using masks
created with a developed algorithm. To create the masks, the following method
was used: a thresholded was applied to the images where intensities below -300
Houns eld units were set as background, the pixel values were normalized to
have an intensity range from 0 to 255, and resulting images were passed through
a binary thresholding process with a threshold value of 20. Using scikit-image1,
small holes and small objects were removed, using methods with the same name
and parameterized with minimum size of 100 and connectivity of 4. Next, the
two methods were reapplied but with a minimum size of 1000. The result is the
desired masks.</p>
        <p>
          Obtained masks were highly similar compared to those provided to the
participants [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. Dice's coe cient, which is scaled from 0 to 1 with 1 corresponding
to image equality, was computed to assess the similarity between created masks
and the original provided masks, with a global average value of 0.9755 being
obtained. Regarding the test dataset, provided masks were used to ensure that
test data to feed the 3D-CNN was not tampered.
        </p>
        <p>After this, the resulting masked volumes were reshaped to comply with the
NHWC channel ordering (number of samples x height x width x channels) used
in CNNs. In our case, the number of samples corresponds to the number of CT
slices. Next, each CT slice was resized to dimensions of 256x256 pixels.</p>
        <p>The resulting volumes were resized, regarding the number of slices, so that
all volumes had the same number of slices. This was achieved by padding the
top and bottom of each volume, resulting in a nal volume with xed size (real
data in the center, and padding in the extremities). Finally, data was normalized
to have zero mean.</p>
        <p>For each subtask, processed datasets were saved in HDF5 les resulting in
two HDF5 les per task with the train and test sets.</p>
      </sec>
      <sec id="sec-2-2">
        <title>Deep Learning</title>
        <p>As expressed, we opted by a 3D-CNN model for the DL model stage. The model
was implemented with TensorFlow [8] version 1.0.0 with support for GPU, which
massively increases the speed and e ciency of training and developing models
such as neural networks. Moreover, TensorBoard was used during the
development of the 3D-CNN model for debugging and optimization purposes. Some
additional functions needed for the models' development were imported from
TFLearn2 (v0.3), a DL library that provides a higher-level API to TensorFlow.</p>
        <sec id="sec-2-2-1">
          <title>1 http://scikit-image.org</title>
        </sec>
        <sec id="sec-2-2-2">
          <title>2 TFLearn: http://t earn.org</title>
          <p>The 3D-CNN model training ran on an Ubuntu server machine equipped with
an NVIDIA Tesla K80 GPU accelerator.</p>
          <p>Regarding the DL model itself, Figure 2 presents an overview of the built
model with the respective composition of each layer. The decision to use a
3DCNN model with seven convolutional layers and two fully connected layers was
empirical.
3D Convolution
3D Max Pooling
Batch Normalization
Leaky ReLU
Layer
3D Convolution
3D Global Avg Pool
Leaky ReLU
Dropout
Layer</p>
          <p>Fully-Connected
Leaky ReLU
Dropout
Layer</p>
          <p>Fully-Connected
Layer</p>
          <p>Result
Input</p>
          <p>Literature supports that deeper models can be more powerful than shallow
ones, as the former can learn how to represent high-level abstractions, presenting
particular interest for the elds of vision, language and other AI-level tasks [9].
However, it is also known that deeper models are more di cult to train due to
problems such as the vanishing gradient problem, where initial layers learn at
slower speeds than nal layers. Naturally, the deeper the network, the more prone
it is to the vanishing gradient problem [10]. Moreover, deeper models demand
bigger compute power, which is a very signi cant overhead. Thus, bearing in
mind the associated implications of creating a deep neural network, and the
existing limited compute power, it was decided to build a network with a small
number of layers.</p>
          <p>As it is possible to observe in Figure 2, the network's rst six layers share the
same structure (but not the hyperparameters). In these six layers, the incoming
tensor is passed through a sequence of 3D convolution, 3D max pooling, batch
normalization and non-linear activation function.</p>
          <p>Batch normalization is very important as it addresses a phenomenon called
internal covariate shift, which slows down the training of neural networks [11].
Concerning the activation function, since the sigmoid activation function can
cause problems when training deep neural networks [12], a variation of the
rectied linear unit (ReLU) { the leaky ReLU { which can lead to better performances
was used in this neural network [13].</p>
          <p>Over tting is other serious concern in neural networks, specially when dealing
with medical imaging datasets which frequently consist of reduced amounts of
data, with unbalanced distributions. For that reason, dropout [14], a regularizer
used to reduce over tting in neural networks, was used in our model. However,
it was only applied to the fully-connected part of the network as convolutional
layers have considerable inbuilt resistance to over tting [15]. Also, L2
regularization was used in each convolutional layer to reduce model over tting, and the
last Fully-Connected layer has a softmax activation function.</p>
          <p>The described 3D-CNN was used for both subtasks, though with di erent
hyperparameters due to the ne-tuning procedure performed for each subtask.
All Leaky ReLUs were used with the leaking coe cient = 0:1 and Dropout
with a drop probability of 0:5 for both subtasks. Table 1 summarizes the
remaining hyperparameters for the models' layers. It should be noted that the
hyperparameters were de ned with compute power constraints in mind. All weights
were initialized as described in [16].</p>
          <p>Concerning data handling, the training dataset was split into 80/20 parcels,
for training and validation splits respectively. Data distribution had moderately
balanced classes for the MDR detection subtask, whereas for the TB type subtask
a less balanced dataset was provided. For each subtask, data was split taking
into account class distributions, in order to ensure the same class distribution in
training and validation splits. Even though the network was prepared to work
with K-fold cross validation, due to time constraints and the inherent nature of
the training process of a neural network, the network was validated o ine using
a single combination of the 80/20 split.</p>
          <p>2
5
7
5
3
2
3
2
3
2
2
2
1</p>
          <p>L2
0.01
0.001
0.002
0.002
0.002
0.003
20
15
15
15
32
64
128
128
Num
Classes
11
7
5
5
9
5
7
5
5
2
3
2
1
Layer
Conv1
Max1
Conv2
Max2
Conv3
Max3
Conv4
Max4
Conv5
Max5
Conv6
Max6
Conv7
FC1
FC2
35
60
60
60
92
92
128
128
Num
Classes
11
5
7
5
5
3
5
3
5
2
3
2
1
3
2
3
2
3
2
7
2
3
2
2
2
1</p>
          <p>L2
0.001
0.001
0.001
0.001
0.001
0.001</p>
          <p>Furthermore, the model was fed with mini-batches of data containing
complete CT scans, where each sample is one of the CT volumes being forwarded
through the net. Since this type of network is demanding in terms of memory
and computational cost, and aiming to enable the use of bigger batch sizes, each
pre-processed volume was cropped into a xed smaller number of slices,
corresponding to the size of the smallest volume in the original dataset. As expressed,
this cropping method extracts data from the center of each CT volume.</p>
          <p>In order to prevent the network from learning a given data sequence/order,
data splits were shu ed in each epoch, prior to being fed to the model. Finally, a
group of four metrics was used to assess model performance, consisting of: cross
entropy, accuracy, precision, and recall.
3</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Submitted Runs and Results</title>
      <p>A single run was submitted for each subtask of the ImageCLEFtuberculosis task,
with the results and respective neural network con gurations being discussed in
this section.</p>
      <p>In both subtasks the neural network models were trained using an Adam
optimizer [17] for stochastic optimization, with the following settings being used:
= LearningRate, 1 = 0:9, 2 = 0:999 and = 10 8. Table 2 summarizes
some of the hyperparameters used in order to train the neural network for each
subtask. All hyperparameters were de ned empirically.</p>
      <p>As previously mentioned, the graphics card used to accelerate the training
of the neural network was an NVIDIA Tesla K80, which possesses two separate
GPUs. Due to the use of the graphics card for other tasks, the MDR
detection model was trained using a single GPU whereas the TB type model was
trained using both GPUs. Therefore, and as shown in Table 2, it was possible
to signi cantly increase the batch size for the TB type network.</p>
      <p>Learning rate was reduced by a fraction of 5 percent of its value after 10 and
15 epochs for subtask 1 and subtask 2, respectively. Validation was performed
in intervals of 3 epochs for the MDR detection's model, and in intervals of 2
epochs for the TB type's model.</p>
      <p>The best results obtained for each subtask during the validation phase are
shown in Table 3. In MDR detection, which is a two class problem, the trained
model favors the retrieval of the most frequent class but struggles to detect
the less frequent and more relevant class, leading to a substantially lower recall
comparatively to obtained accuracy and precision.</p>
      <p>In the TB type task, a multi-class problem ( ve classes), it is possible to see
that the tuned model attained a lower accuracy, while keeping slightly similar
precision and improved recall. The impact of having a higher number of classes,
combined with a less balanced dataset for this task had a repercussion on the
validation accuracy which was signi cantly lower than in the MDR detection
task. Such accuracy value demonstrates that the neural network had di
culties in identifying the classes in data, which explains why some classes had no
occurrence registered in the validation dataset.</p>
      <p>Regarding the testing phase, in the MDR detection subtask contestants had
to submit the probability of each patient having MDR, whereas for the TB
type task submissions had to contain the expected TB type for each TB
patient. Model performance was assessed with di erent metrics for each subtask:
in MDR detection, performance was measured with Accuracy and Area Under
the Curve (AUC) obtained from the ROC-curves produced with the submitted
probabilities; in TB type classi cation Accuracy and Cohen's Kappa were the
selected metrics. Table 4 presents test results both for the submitted run and
for the best run in each subtask.</p>
      <p>
        The list of test results for the two subtasks comprised in ImageCLEF's
tuberculosis task [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] clearly demonstrates the high di culty associated with this
challenge's proposition. On the one hand, submitted runs performed worse in
each subtask than the remaining entries. On the other hand, for the MDR
detection subtask, the top ranking model had an accuracy just slightly over 50%
whilst our model's test accuracy was nearly 47% (the best overall accuracy was
56.8%). Concerning AUC, our model scored lower than the top ranking entry by
a bigger margin.
      </p>
      <p>For the TB type subtask, test results were in general worse comparatively
to results of the MDR detection subtask. In this subtask, the top ranking entry
had an accuracy of 40% compared to our model's 24.3%, and a Cohen's Kappa
of 0.24 compared to the marginal value of 0.02 obtained by our model.</p>
      <p>Aside from the comparison with other models' performance, it is noticeable
that for subtask 1 our model had lower accuracy in the test phase (46.5%) than
in the validation phase (55%), whereas for subtask 2 the opposite occurred with
test accuracy (24%) being higher than validation accuracy (17.4%). For the rst
part, it is very likely that the model su ered over tting to the training dataset
(a common issue when dealing with medical imaging datasets), and testing
performance su ered a signi cant impact from that. For the second part, there
exists the possibility of having a test dataset less balanced, regarding class
distribution, than the training dataset. By having a class distribution more skewed
towards the more frequent classes, our model can attain higher accuracy scores
than during the validation process.</p>
      <p>
        In spite of our models' poor performance in general, the nal
ImageCLEFtuberculosis result list [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] shows that there are other entries with comparable
performance. Fine-tuning a model is a slow, thorough process that should be
methodical. In our approach, the search for the best hyperparameters was
empirical and not extensive enough due to limitations in terms of available time.
There is much con dence that there exists a big margin for progress and
improvement in our work, provided there is more time to better train the models,
and correctly ne-tune them.
4
      </p>
    </sec>
    <sec id="sec-4">
      <title>Conclusion and Future Work</title>
      <p>The ImageCLEFtuberculosis task is a challenge that encompasses the classi
cation problem on medical images. This task was divided into two subtasks: Drug
Resistance Detection and TB classi cation. In the rst subtask the objective was
to assess the probability of a TB patient having a resistant form of tuberculosis,
whereas on the second one the goal was to classify the TB type from a pool of
ve possible types.</p>
      <p>In this paper we presented two separate runs that were submitted for the
two subtasks. In both subtasks, provided images were pre-processed for this
challenge. Although the test results of our submitted runs for both subtasks
were low (46.5% accuracy, AUC of 0.46 and 24% accuracy, Cohen's Kappa of
0.022, respectively), the majority of the submitted runs behaved in a similar
way, since the di erences in terms of accuracy between the best submitted run
and our own were of 5% and 15% for the MDR detection and TB type subtask,
respectively. As a side note, it is interesting to notice in the list of submissions
that various entries used DL approaches to tackle this challenge, which shows
that DL is an area that holds great promise.</p>
      <p>Since over tting was an e ective reality during the development of the neural
network models, in the future we hope to evaluate the impact of techniques
such as data augmentation and weight normalization on our models' results.
Furthermore, running the model with K-fold cross-validation and performing an
ensemble of the resulting networks could further improve our results.</p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgments</title>
      <p>This work is nanced by the ERDF - European Regional Development Fund
through the Operational Programme for Competitiveness and
Internationalisation - COMPETE 2020 Programme, and by National Funds through the FCT
Fundac~ao para a Ci^encia e a Tecnologia. Jo~ao Figueira Silva is funded by the
research grant of PTDC/EEI-ESS/6815/2014 project and Jorge Miguel Silva
is funded by the research grant of CMUP-ERI/ICT/0028/2014-SCREEN-DR
project. Eduardo Pinho also was funded by the FCT under the grant PD/BD/
105806/2014.
8. Abadi, M., Agarwal, A., Barham, P., Brevdo, E., Chen, Z., Citro, C., Corrado,
G.S., Davis, A., Dean, J., Devin, M., Ghemawat, S., Goodfellow, I., Harp, A.,
Irving, G., Isard, M., Jia, Y., Jozefowicz, R., Kaiser, L., Kudlur, M., Levenberg,
J., Mane, D., Monga, R., Moore, S., Murray, D., Olah, C., Schuster, M., Shlens, J.,
Steiner, B., Sutskever, I., Talwar, K., Tucker, P., Vanhoucke, V., Vasudevan, V.,
Viegas, F., Vinyals, O., Warden, P., Wattenberg, M., Wicke, M., Yu, Y., Zheng, X.:
TensorFlow: Large-scale machine learning on heterogeneous distributed systems.
(2016)
9. Bengio, Y., Yoshua: Learning Deep Architectures for AI. Foundations and</p>
      <p>Trends R in Machine Learning 2(1) (2009) 1{127
10. Hochreiter, S., Bengio, Y., Frasconi, P., Schmidhuber, J.: Gradient Flow in
Recurrent Nets: the Di culty of Learning Long-Term Dependencies. In: Field Guide to
Dynamical Recurrent Networks. IEEE Press (2001)
11. Io e, S., Szegedy, C.: Batch Normalization: Accelerating Deep Network Training
by Reducing Internal Covariate Shift. CoRR abs/1502.03167 (2015)
12. Glorot, X., Bengio, Y.: Understanding the di culty of training deep feedforward
neural networks. In: Proceedings of the International Conference on Arti cial
Intelligence and Statistics (AISTATS10). Society for Arti cial Intelligence and
Statistics. (2010)
13. Maas, A.L., Hannun, A.Y., Ng, A.Y.: Recti er Nonlinearities Improve Neural
Network Acoustic Models. In: ICML Workshop on Deep Learning for Audio, Speech
and Language Processing. Volume 30. (2013)
14. Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.:
Dropout: A Simple Way to Prevent Neural Networks from Over tting. Journal
of Machine Learning Research 15 (2014) 1929{1958
15. Nielsen, M.A.: Neural Networks and Deep Learning (2015)
16. He, K., Zhang, X., Ren, S., Sun, J.: Delving deep into recti ers: Surpassing
humanlevel performance on imagenet classi cation. In: Proceedings of the IEEE
International Conference on Computer Vision. (2015) 1026{1034
17. Kingma, D.P., Ba, J.: Adam: A method for stochastic optimization. CoRR
abs/1412.6980 (2014)</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>Dicente</given-names>
            <surname>Cid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            ,
            <surname>Kalinovsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Liauchuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            ,
            <surname>Kovalev</surname>
          </string-name>
          ,
          <string-name>
            <surname>V.</surname>
          </string-name>
          , , Muller, H.:
          <article-title>Overview of ImageCLEFtuberculosis 2017 - Predicting Tuberculosis Type and Drug Resistances</article-title>
          . CLEF working notes,
          <source>CEUR</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Ionescu</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          , Muller, H.,
          <string-name>
            <surname>Villegas</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Arenas</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Boato</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dang-Nguyen</surname>
            ,
            <given-names>D.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dicente Cid</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Eickho</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garcia Seco de Herrera</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gurrin</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Islam</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kovalev</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liauchuk</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mothe</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Piras</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Riegler</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schwall</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Overview of ImageCLEF 2017: Information extraction from images</article-title>
          .
          <source>In: Experimental IR Meets Multilinguality, Multimodality, and Interaction 8th International Conference of the CLEF Association</source>
          ,
          <string-name>
            <surname>CLEF</surname>
          </string-name>
          <year>2017</year>
          . Volume
          <volume>10456</volume>
          of Lecture Notes in Computer Science., Dublin, Ireland, Springer (September
          <volume>11</volume>
          -14
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Russakovsky</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Deng</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Su</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Krause</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Satheesh</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , Ma,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            ,
            <surname>Karpathy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Khosla</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Bernstein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Berg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.C.</given-names>
            ,
            <surname>Fei-Fei</surname>
          </string-name>
          ,
          <string-name>
            <surname>L.</surname>
          </string-name>
          :
          <article-title>ImageNet Large Scale Visual Recognition Challenge</article-title>
          .
          <source>International Journal of Computer Vision</source>
          (IJCV)
          <volume>115</volume>
          (
          <issue>3</issue>
          ) (
          <year>2015</year>
          )
          <volume>211</volume>
          {
          <fpage>252</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Bengio</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Courville</surname>
            ,
            <given-names>A.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vincent</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Representation Learning: A Review and New Perspectives</article-title>
          .
          <source>CoRR</source>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Hinton</surname>
            ,
            <given-names>G.E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Salakhutdinov</surname>
            ,
            <given-names>R.R.</given-names>
          </string-name>
          :
          <article-title>Reducing the Dimensionality of Data with Neural Networks</article-title>
          .
          <source>Science</source>
          <volume>313</volume>
          (
          <issue>5786</issue>
          ) (
          <year>2006</year>
          )
          <volume>504</volume>
          {
          <fpage>507</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Ravi</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wong</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Deligianni</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Berthelot</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Andreu-Perez</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lo</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>G.Z.</given-names>
          </string-name>
          :
          <article-title>Deep Learning for Health Informatics. Biomedical and Health Informatics</article-title>
          ,
          <source>IEEE Journal of 21</source>
          (
          <issue>1</issue>
          ) (jan
          <year>2017</year>
          )
          <volume>4</volume>
          {
          <fpage>21</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>Dicente</given-names>
            <surname>Cid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            ,
            <surname>del Toro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.A.</given-names>
            ,
            <surname>Depeursinge</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.</surname>
          </string-name>
          , Muller, H.:
          <article-title>E cient and fully automatic segmentation of the lungs in CT volumes</article-title>
          .
          <source>In: Proceedings of the VISCERAL Anatomy Grand Challenge at the 2015 IEEE ISBI. CEUR Workshop Proceedings</source>
          , CEUR-WS (
          <year>2015</year>
          )
          <volume>31</volume>
          {
          <fpage>35</fpage>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>