<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>ImageCLEF 2019: Deep Learning for Tuberculosis CT Image Analysis</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Abdelkader Hamadi[</string-name>
          <email>abdelkader.hamadi@univ-mosta.dz</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Noreddine Belhadj Cheikh</string-name>
          <email>noreddine.belhadjcheikh@univ-mosta.dz</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yamina Zouatine</string-name>
          <email>zouatineyamina@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Si Mohamed Bekkai Menad</string-name>
          <email>bekkai.menad@univ-mosta.dz</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mohamed Redha Djebbara</string-name>
          <email>redha.djebbara@univ-mosta.dz</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Abdelhamid Ibn Badis Mostaganem Faculty of Exact Sciences and Computer Science Mathematics and Computer Science Department Mostaganem</institution>
          ,
          <country country="DZ">Algeria</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In this article, we present the methodologies used in our participation in the two subtasks of the ImageCLEF 2019 Tuberculosis Task (SVR and CTR). Our contributions are essentially based on deep learning and other machine learning techniques. In addition to the use of deep learners, semantic descriptors are tested to represent patients CT scans. These features are extracted after a rst learning step. Our submissions on the test corpus reached AUC value of about 65% in the SVR task and an average AUC value of about 63% in CTR. These results o ered us the seventh and the eighth places in SVR and CTR, respectively. We believe that our contributions could be further improved and might give better results if they applied properly and in an optimized way.</p>
      </abstract>
      <kwd-group>
        <kwd>ImageCLEF Tuberculosis Task Deep Learning CT Image Tuberculosis CT Image Classi cation Tuberculosis Severity Scoring CT Report</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Tuberculosis (TB) is a deadly disease. Its early diagnosis can give the necessary
treatment and prevent the death of patients. The technological advancement
especially in the eld of arti cial intelligence and precisely supervised learning
opens the door for researchers to study the possibility of an automatic diagnosis.
This would speed up the process and lower its cost. Several researchers have
invested their e orts in recent years, especially within the medical image analysis
community. In fact, a task dedicated to this disease had been adopted as part of
the ImageCLEF evaluation campaign in its editions of the three last years [
        <xref ref-type="bibr" rid="ref6 ref7 ref8">6, 8,
7</xref>
        ]. In this task, the objective is to analyze automatically the 3D CT images of
TB patients to detect semantic information: the type of Tuberculosis, the degree
of severity of the disease, information related to the state of the lungs, etc. In
ImageCLEF 2019 two sub-tasks of the main task, \ImageCLEFmed
Tuberculosis" are considered: Severity Scoring (SVR) and CT Report (CTR). In the rst
task, the goal is to deduce automatically from a CT image whether a TB case
is severe or not. In the second one, the problematic consists of generating an
automatic report that includes the following information in binary form (0 or
1): Left lung a ected, right lung a ected, presence of calci cations, presence of
caverns, pleurisy, lung capacity decrease. based solely on the CT image. We can
summarize the objectives of the Tuberculosis task through the following points:
{ Helping medical doctors in the diagnosis and determining the state of the
patient through image processing techniques;
{ Predicting quickly the TB severity degree to make quick decisions and give
e ective treatments;
{ Assist doctors and medical o cers to have accurate details about the
patient's lung condition by providing a report summarizing information
describing the state of the lungs.
      </p>
      <p>
        We present in the following section our work that had been made in the
context of our participation in the two sub-tasks of ImageCLEF 2019
Tuberculosis task: Tuberculosis Severity Scoring (SVR) and Tuberculosis CT Report
(CTR) [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>The remainder of this article is organized as follows. Section 2 describes the
two tasks in which we had participated. In section 3, we present our contribution
by detailing the system deployed to perform our submissions. Section 4 details
our experimental protocols we used to generate our predictions. We present and
analyze in the same section the results obtained. We make our conclusions in
the last section by presenting potential perspectives and future works.
2</p>
      <p>
        Participation to ImageCLEF 2019
ImageCLEF 2019 [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] is an evaluation campaign that is being organized as part
of the CLEF initiative labs. This campaign o ers several research tasks that
welcome participation from teams around the world. For the 2019 edition,
ImageCLEF organises four main tasks: ImageCLEFcoral, ImageCLEFlifelog,
ImageCLEFmedical and ImageCLEFsecurity. In this work, we focus on the
Tuberculosis task that takes part in the ImageCLEFmedical challenge. ImageCLEFmed
Tuberculosis task includes two sub-tasks: Severity Scoring (SVR) and CT Report
(CTR) that we describe in the following.
2.1
      </p>
    </sec>
    <sec id="sec-2">
      <title>SVR and CTR Tasks description</title>
      <p>In this paper, we focus on our participation in the SVR and the CTR sub-tasks.
The main objective of these two challenges is the automatic analysis of
Tuberculosis CT scans. In both tasks, the same dataset is used, one corpus for training
and another one for testing. The data is provided as 3D CT scans. All the CT
images are stored in NIFTI le format with \nii.gz" extension le (gzipped .nii
les). For each of the three dimensions of the CT image, we nd a number of
slices varying according to the dimension considered (512 images for the Y and
X dimensions, from 40 to 250 images for the Z dimension). Each slice has a size
of about 512 128 pixels for the X and Y dimensions and 512 512 pixels for the
Z dimension.</p>
      <p>A training collection is provided at the beginning of the task with its
groundtruth (labels of samples). Participants prepare and train their systems on this
dataset. A test collection is provided at a later date. Participants interrogate
their system and submit their predictions to the organizers' committee. An
evaluation is performed by the latter to compare the performance of the participants'
predictions submissions.</p>
      <p>SVR task aims to predict the degree of severity of TB cases. Given a CT scan
of TB patient, the main goal is to predict the severity of his illness based on his
3D CT scan. The degree of severity is modeled according to 5 discrete values:
from 1 (\critical/very bad") to 5 (\very good"). The score value is simpli ed so
that values 1, 2 and 3 correspond to \high severity" class, and values 4 and 5
correspond to \low severity".</p>
      <p>The classi cation problem is evaluated using two measures: 1) Area Under
ROC-curve (AUC) and 2) Accuracy.</p>
      <p>CT Report task has as objective to automatically generate a report based
on the patient's CT scan. This report should include the following six pieces of
information in the binary form (0 or 1):
1. Is the left lung a ected?
2. Is the right lung a ected?
3. The presence of calci cations;
4. The presence of caverns;
5. The presence of pleurisy;
6. The lung capacity decrease.</p>
      <p>This task is considered as a multilabel classi cation problem (6 binary
ndings). The ranking of this task is done rst by average AUC and then by min
AUC (both over the 6 CT ndings).</p>
      <sec id="sec-2-1">
        <title>Our contributions</title>
        <p>We proposed to use the system presented in Figure 1. The latter goes through
two essential steps: input data pre-processing and training a classi cation model.
A third optional step is added in order to improve the performance of the rst
learning step. The latter includes a second learning stage by using a recurrent
neural network (LSTM) or by generating semantic features and exploiting them
through a learner or a deep learner. We will detail our proposed system in the
following.</p>
        <sec id="sec-2-1-1">
          <title>Nifti</title>
        </sec>
        <sec id="sec-2-1-2">
          <title>CT scans</title>
        </sec>
        <sec id="sec-2-1-3">
          <title>Data preprocessing</title>
        </sec>
        <sec id="sec-2-1-4">
          <title>Generating</title>
        </sec>
        <sec id="sec-2-1-5">
          <title>Features</title>
          <p>of slices</p>
        </sec>
        <sec id="sec-2-1-6">
          <title>Generating</title>
        </sec>
        <sec id="sec-2-1-7">
          <title>Semantic</title>
          <p>descriptors
labels
Slices</p>
        </sec>
        <sec id="sec-2-1-8">
          <title>LSTM</title>
        </sec>
        <sec id="sec-2-1-9">
          <title>Learner or</title>
        </sec>
        <sec id="sec-2-1-10">
          <title>Deep Learner</title>
          <p>We remind that in both tasks, 3D CT scans are provided in compressed Nifti
format. Firstly, we decompressed the les and extracted the slices. In the end,
we got three sets of slices corresponding to the three dimensions of the 3D
image. For each dimension and for each Nifti image we obtained a number of slices
ranging according to the dimension considered (512 images for the Y and X
dimensions, from 40 to 250 images for the Z dimension).</p>
          <p>The visual content of the images that were extracted from the di erent
dimensions is not similar. Indeed, the images of each dimension are taken from a
di erent angle of view. We noticed from our experiments that the slices of the
-Z- dimension give better results compared to the two others (X and Y). This
remark concerns our proposed approaches. This is why we used in our work the
Z-dimension. However, all steps can be applied to slices of any of the three
dimensions.</p>
          <p>After choosing the dimension to consider, we propose to lter the slices of
each patient. Indeed, we can notice that many slices do not necessarily contain
relevant information that could help to classify the samples. This is why we
added a step to lter and select a number of slices per patient. For this, we
propose two ltering approaches:</p>
          <p>
            Automatic supervised ltering: In this approach, we select a set of
patients from each of the considered classes (the ve degrees of severity for the
SVR task). Then, a professional radiologist selects for each patient, the slices
likely to contain relevant information indicating the presence of Tuberculosis.
The resulting set of slices constitutes a ltering group. Given a new patient, we
compare each of its slices to the ltering group by calculating a distance measure:
a weighted sum of distances between the slice and those of the ltering group.
This comparison can be done through a direct pixel-wise comparison. In our
experiments we used the \Structural Similarity" as distance [
            <xref ref-type="bibr" rid="ref4">4</xref>
            ]. Unfortunately, we
could not do a thorough to choose a better distance for lack of time. We selected
at the end N slices that are judged to be the most similar to the ltering group.
So, at the end, each patient is represented by the N ltered slices instead of all
its extracted images. We think that this would reduce the noise introduced by
the consideration of all slices. We tested in our contributions the value N=10.
          </p>
          <p>CT scans
Nifti format
(.nii.gz)</p>
          <p>Selecting a
dimension</p>
          <p>Z</p>
          <p>Extracting</p>
          <p>slices
Converting Nifti
to JPG</p>
          <p>Image slices
3 Dimensions
X</p>
          <p>Y</p>
          <p>Z</p>
          <p>Filtering slices using;
Automatic supervised filtering</p>
          <p>or
Automatic unsupervised filtering</p>
          <p>N selected slices per</p>
          <p>CT image / patient</p>
          <p>Automatic unsupervised ltering: We noticed that there is usually a
maximum of 50/60 slices visually informative. Since the slices are ordered, the
most informative slices are usually at the center of the list. We propose then to
keep only the N middle ones. This is not optimal but we opted for this choice for
a fully automatic and unsupervised approach. This choice can be improved by
performing manual ltering with the intervention of a human expert, preferably
with medical skills on TB disease.</p>
          <p>
            Deep learning model for CT image classi cation ( rst learning
step)
As a deep learner, we chose to use Resnet-50 architecture because of its good
results in the context of the same problematic in last Tuberculosis task
editions [
            <xref ref-type="bibr" rid="ref13">13</xref>
            ]. On the other hand, we developed a model that we called \LungNet".
We present more details about this deep learner in the following section. The
outputs of the deep learners deployed are considered as initial results. We
exploited then these outputs to generate: 1) semantic features of a patient that are
used to reclassify the samples, and 2) features of slices organized in a sequence
format that are fed to LSTM input as described in section 3.5.
          </p>
          <p>
            LungNet Deep Learner: We proposed and developed our deep learner
architecture for CT Image Analysis that we called \LungNet". The input to the
latter is an RGB image of size 119x119, followed by ve convolutional layers and
two fully connected layers. Initially, input data were in nifty format. Slices of
the CT scans are 1-channel gray-level images. However, we extracted the slices
using med2image tool [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ]. This software converts the slices to jpeg format. To
avoid introducing noise by using this extraction method, we can do better by
reading directly image pixels values using Niftilib library for python that was
suggested by the task organizers. The idea behind using med2image to extract
slices is that we planned to lter the slices by a medical expert intervention.
This process required the slices to be in a format easily visible to the expert.
          </p>
          <p>After each convolutional layer \relu" activation is applied followed by a local
normalization and MaxPooling. The rst, the second and the third convolution
blocks have dropout layers to reduce over tting. The sigmoid activation
function is applied to the output layer in order to predict values in the range of 0 to 1.</p>
          <p>
            Figure 3 illustrates the architecture of the Lungnet model.
We implemented the method of semantic descriptors extraction described in [
            <xref ref-type="bibr" rid="ref9">9</xref>
            ]
with slight di erences. After slices extraction and ltering, we generated a single
descriptor per patient to exploit it through a transfer learning process. The
results of SGEast [
            <xref ref-type="bibr" rid="ref13">13</xref>
            ] and even other teams in the same task of ImageCLEF 2017
proved the e ciency of this approach [
            <xref ref-type="bibr" rid="ref13 ref9">13, 9</xref>
            ].
          </p>
          <p>So, we chose to exploit the probabilities that were predicted by a deep learner
trained on a set of slices. If K is the number of classes considered, these
predictions typically correspond to the K predicted probability values for the K
classes (For SVR Task, K = 5: the ve severity degrees). We obtain then for
each slice K values corresponding to the probabilities of the K considered classes.</p>
          <p>Furthermore, K sub-descriptors are generated: D1, D2, D3, D4, ... Dk. Each
sub-descriptor Di contains the predicted probabilities for the class i for all the
slices of the patient. A nal semantic descriptor is constructed by concatenating
the K sub-descriptors. Figure 4 details the whole process of the semantic features
extraction for one patient.
Slices</p>
          <p>
            Learning a classi cation model based on semantic features
(second learning step)
We propose to exploit the semantic descriptors of patients described previously.
Any approach of supervised classi cation can be applied as shown in gure 5.
We tested in our experiments SVM as supervised classi er. However, Random
Forests and bagging of Random Forests have shown good results in the context
of the same problematic [
            <xref ref-type="bibr" rid="ref9">9</xref>
            ].
          </p>
          <p>
            We recommend some ideas for this step:
{ To use a deep learner having as input the semantic descriptors of patients
and the labels of patients. As an alternative, it would be interesting to use a
bagging method that collaborates several learners and sub-samples the train
collection. This would lead to better results as presented in [
            <xref ref-type="bibr" rid="ref9">9</xref>
            ];
{ To apply samples selection and data augmentation.
Train –corpus
CT images
CT image
of a test
patient
n
o
i
t
c
a
r
t
x
e
s
e
r
u
t
a
e
F
c
i
t
n
a
m
e
S
          </p>
          <p>Semantic descriptors / Train corpus</p>
          <p>Semantic descriptor 1
Semantic d.escriptor 2
.
.</p>
          <p>.</p>
          <p>
            Semantic descriptor n
As each patient is described by a sequence of slices, it is interesting to test the
LSTM (Long Short-Term Memory) [
            <xref ref-type="bibr" rid="ref10">10</xref>
            ] recurrent neural network that is suitable
for such data type. However, it is not recommended to apply LSTM on slices
as input. Extracting features from slices using deep learner and pushing them
to LSTM seems to be a good alternative. We propose to describe each slice by
a feature of size equal to the number of the considered classes ( ve classes for
SVR task). This feature is composed of the ve values corresponding to the
probabilities of the considered classes. These values are obtained through a deep
learning stage. After generating these features, they are fed to an LSTM neural
network by considering the ordered set of slices of each patient as a sequence.
Figure 6 describes the whole process.
We describe in the following sections our main runs submitted to the SVR and
CTR tasks.
          </p>
          <p>
            We used in our experimental work the following tools:
{ med2image [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ] for the conversion of nifti medical images to the classic Jpeg
format;
{ Tensor ow frawework [
            <xref ref-type="bibr" rid="ref3">3</xref>
            ] and Keras library [
            <xref ref-type="bibr" rid="ref5">5</xref>
            ] for deep learning;
{ scikit-learn [
            <xref ref-type="bibr" rid="ref12">12</xref>
            ] library for testing several machine learning techniques.
          </p>
          <p>We chose to use slices of the -Z- dimension because our experiments showed
that they are more suitable than those of the two other ones and got better
results.</p>
          <p>Dataset: The dataset used in the SVR task includes chest CT scans of TB
patients along with some metadata that describe a set of 19 classes. 2 classes
concern the SVR task, six other classes concern CTR task. The other values are
considered as additional information regarding to the patients. They could be
used as contextual information. Table 1 summarizes the number of CT scans for
train and test collections.</p>
          <p>The same dataset is given for the CTR task. The samples are labeled
regarding to seven main target classes:
1. Target classes for SVR Task:
(a) SVR severity (binary class: HIGH and LOW). Another label called md Severity
is given (Five discrete values ranging from 1 to 5). We remind that
values of md Severity (1, 2 and 3) belong to the \HIGH" Severity case. The
other two values (4 and 5) correspond to the \LOW" Severity.
2. Target classes for CTR Task (binary classes):
(a) Left lung a ected;
(b) Right lung a ected;
(c) Presence of calci cations;
(d) Presence of caverns;
(e) Presence of pleurisy;
(f) Lung capacity decrease.
4.1</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>SVR task</title>
      <p>Experimental protocol: We used the train collection provided by the
organizers and we split it into two sub-collections: training and validation sets. We
nally submitted three main runs. The other submissions concern some tested
approaches that we could not optimize and nalize correctly because of lack of
time:
{ SVR FSEI resnet50 run3: results of ResNet-50 trained on 50% of
training data. Each patient was represented by 50 slices ltered using the
automatic unsupervised ltering approach that was described in section 3.1.
The slices were adapted by resizing them directly using the Python Imaging
Library (PIL). The input images of Resnet50 are of size 199 199;
{ SVR FSEI lungnet run2: results of LungNet deep learner trained on 80%
of data. Each patient was represented by 10 slices ltered using the
automatic supervised ltering approach that was described in section 3.1;
{ SVR FSEI lstm run8: results of LSTM exploiting outputs of Lungnet
deep learner. Each patient was represented by 50 slices ltered using the
automatic unsupervised ltering approach. So, a sequence for the LSTM
learner is composed of the 50 features representing the 50 slices of the
patient.</p>
      <p>We considered for each run a hierarchical classi cation problem. Firstly, we
classi ed the samples in the 5 classes corresponding to the ve degrees of severity.
Secondly, We deduced for each patient its predicted class using a majority vote
on the predicted labels of all slices. Finally, the class predicted in the previous
step is transformed to a binary value corresponding to the SVR Severity class
(HIGH if predicted class 2 f1; 2; 3g and LOW if not).</p>
      <p>
        Our tools and scripts used in our experiments are accessible in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
Results: Table 2 shows the results in terms of AUC and accuracy obtained
by our runs on the evaluation performed by the ImageCLEF committee on test
collection.
      </p>
      <p>We can see that SVR FSEI resnet50 run3 got the best performance
followed by SVR FSEI lstm run8. These two runs were ranked 22th and 25th
out of 54 submissions.</p>
      <p>We note that for SVR FSEI lungnet run2 patients were represented by
10 slices (50 slices for the two other runs), it would be interesting to see the
performance of the Lungnet model after training it on 50 slices per patient in
order to make a detailed comparison with the other two runs.</p>
      <p>Figures 7 and 8 describe the results and ranking of all submissions of SVR
task in terms of AUC and accuracy, respectively.</p>
      <p>Although the results achieved by our submissions are not well ranked
compared to those of the top of the list, we can notice that several runs belong to the
same teams that had good results, and they probably do not di er too much.
On the other hand, We believe that our models could give better results after a
more advanced data preprocessing including the use of masks, samples selection
and data augmentation.</p>
      <p>Results of imageCLEF2019 SVR task on test collection
R u U U _2</p>
      <p>MM
S s H H n _ _
e e MM</p>
      <p>l
_Mb V V</p>
      <p>R m-S -S
H H u
_ _ r mm
R R _ o o SV se RV RV</p>
      <p>r r
V V V F F n S S
S S SR _ _ tE</p>
      <p>R R s
V V
S S</p>
      <p>A
L
_
R
V
S</p>
      <p>mn
aad sub -LRVAD -LRVAD -SSRVV -SSRVV rr_3nu -LSRVD -LSRVD lts_8n eedoC
t
e
M
R S S
V
S
_ u n
I r _
E
S I_ N
F E N
_ S
R F G</p>
      <p>_ _
V
S</p>
      <p>R R
SV SV</p>
      <p>S R R g R</p>
      <p>V V n V
S S lu S
_
2
n
u
r
_
I
E
S
F
_
R
V
S
ru _ B B ss R _ _ _ E x x
_V bm_D _D le V ta ta t O -a -a</p>
      <p>a f
S e e
0
Accuracy 0.4
Experimental protocol: We trained in a rst step our deep models (Resnet
and Lungnet). Secondly, we generated the semantic descriptors following the
approach described in section 3. We treated the problematic as a multilabel
classi cation problem in the rst learning stage and as a binary classi cation
problem in the second learning stage. We used in the latter SVM as a binary
classi er. We optimized its parameters independently for each target class.</p>
      <p>We submitted three main runs:
1. CTR FSEI run1: results of LungNet trained on 50% of training data. Each
patient was represented by 10 slices ltered using the automatic supervised
ltering approach that was described in section 3.1;
2. CTR FSEI run2 : results of LungNet trained on 70% of training data.</p>
      <p>Each patient was represented by 50 slices ltered using the automatic
unsupervised ltering approach that was described in section 3.1;
3. CTR FSEI run5: SVM using semantic features that are extracted using
Resnet-50. Each patient was represented by 10 slices ltered using the
automatic supervised ltering approach that was described in section 3.1.</p>
      <p>
        Our tools and scripts used in our experiments are accessible in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
Results: Table 3 shows the results (in terms of Average-AUC and Min-AUC)
and ranking obtained by our runs on the evaluation performed by the
ImageCLEF committee on test collection.
      </p>
      <p>
        We can see that our best results were obtained by CTR FSEI run1
followed by CTR FSEI run2. However, we should mention here that we used
the same sub-division of the corpus in two sub-parts (train and validation) for
all CTR target classes, which is not optimal since the distribution of class
values is not the same for the six target classes. This explains the disadvantage of
the run CTR FSEI run5 compared to the other two and also the low value
of Min-AUC for the three runs. We believe that the semantic descriptors
approach might perform better by making more e orts to optimize parameters
or by testing another learner like the Bagging of Random Forests as presented
in [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Considering a multi-label classi er constitutes also an interesting idea to
test.
      </p>
      <p>Results of CTR task</p>
      <sec id="sec-3-1">
        <title>Conclusion and future works</title>
        <p>We have described in this article our contributions to the SVR and CTR
subtasks of ImageCLEFmed 2019 Tuberculosis task. We proposed to use after a
data preprocessing step, a deep learner to classify samples to the target classes.
We used for that, ResNet-50 and proposed our LungNet architecture. Moreover,
we proposed to extract a single semantic descriptor for each CT image/patient
instead of considering all the slices as separate samples. We tested also LSTM as
another alternative. Although our proposals had not been the best, the obtained
results showed that these approaches could be much more e cient and might
give more interesting results if they are applied in an optimized way.</p>
        <p>As perspectives, we plan to adopt data augmentation strategies and learning
samples selection. In addition, we noticed during the sub-sampling of our data
that the deletion or addition of some samples had an impact on the results. On
the other hand, ltering slices in an optimized way is a key idea that could further
improve the system performance. Moreover, we noticed in our experiments that
there is a di erence of precision for each severity class studied which arises
the hypothesis of the classes having varying di culties to be identi ed by the
model. Indeed, some classes are more di cult to identify than others. It is also
an interesting track to study.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1. med2image: https://github.com/fnndsc/med2image. Last check:
          <volume>30</volume>
          /05/
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2. Resnet-
          <volume>50</volume>
          and
          <article-title>lungnet for tuberculosis severity scoring</article-title>
          . mostaganem university at imageclefmed 2019 :
          <article-title>tools to run experiments</article-title>
          . https://github.com/anouar1991/imageCLEFfsei/tree/master/tools/application
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Abadi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Agarwal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Barham</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brevdo</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Citro</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corrado</surname>
            ,
            <given-names>G.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Davis</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dean</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Devin</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ghemawat</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goodfellow</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Harp</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Irving</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Isard</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jia</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jozefowicz</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kaiser</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kudlur</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Levenberg</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mane</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Monga</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moore</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Murray</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Olah</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schuster</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shlens</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Steiner</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sutskever</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Talwar</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tucker</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vanhoucke</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vasudevan</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Viegas</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vinyals</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Warden</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wattenberg</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wicke</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zheng</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          :
          <article-title>TensorFlow: Large-scale machine learning on heterogeneous systems (</article-title>
          <year>2015</year>
          ), http://tensor ow.org/, software available from tensor ow.org
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Brunet</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vrscay</surname>
            ,
            <given-names>E.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          :
          <article-title>On the mathematical properties of the structural similarity index</article-title>
          .
          <source>IEEE Transactions on Image Processing</source>
          <volume>21</volume>
          (
          <issue>4</issue>
          ),
          <volume>1488</volume>
          {1499 (April
          <year>2012</year>
          ). https://doi.org/10.1109/TIP.
          <year>2011</year>
          .2173206
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Chollet</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          , et al.: Keras. https://github.com/fchollet/keras (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>Dicente</given-names>
            <surname>Cid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            ,
            <surname>Kalinovsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Liauchuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            ,
            <surname>Kovalev</surname>
          </string-name>
          ,
          <string-name>
            <surname>V.</surname>
          </string-name>
          , , Muller, H.:
          <article-title>Overview of ImageCLEFtuberculosis 2017 - predicting tuberculosis type and drug resistances</article-title>
          .
          <source>In: CLEF2017 Working Notes. CEUR Workshop Proceedings</source>
          , CEURWS.org &lt;http://ceur-ws.
          <source>org&gt;</source>
          , Dublin,
          <source>Ireland (September</source>
          <volume>11</volume>
          -14
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>Dicente</given-names>
            <surname>Cid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            ,
            <surname>Liauchuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            ,
            <surname>Klimuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Tarasau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Kovalev</surname>
          </string-name>
          ,
          <string-name>
            <surname>V.</surname>
          </string-name>
          , Muller, H.:
          <article-title>Overview of ImageCLEFtuberculosis 2019 - automatic ct-based report generation and tuberculosis severity assessment</article-title>
          .
          <source>In: CLEF2019 Working Notes. CEUR Workshop Proceedings</source>
          , CEUR-WS.org &lt;http://ceur-ws.
          <source>org&gt;</source>
          , Lugano,
          <source>Switzerland (September 9-12</source>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>Dicente</given-names>
            <surname>Cid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            ,
            <surname>Liauchuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            ,
            <surname>Kovalev</surname>
          </string-name>
          ,
          <string-name>
            <surname>V.</surname>
          </string-name>
          , , Muller, H.:
          <article-title>Overview of ImageCLEFtuberculosis 2018 - detecting multi-drug resistance, classifying tuberculosis type, and assessing severity score</article-title>
          .
          <source>In: CLEF2018 Working Notes. CEUR Workshop Proceedings</source>
          , CEUR-WS.org &lt;http://ceur-ws.
          <source>org&gt;</source>
          , Avignon,
          <source>France (September 10- 14</source>
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Hamadi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yagoub</surname>
            ,
            <given-names>D.E.</given-names>
          </string-name>
          :
          <article-title>Imageclef 2018: Semantic descriptors for tuberculosis CT image classi cation</article-title>
          .
          <source>In: Working Notes of CLEF 2018 - Conference and Labs of the Evaluation Forum</source>
          , Avignon, France,
          <source>September 10-14</source>
          ,
          <year>2018</year>
          . (
          <year>2018</year>
          ), http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2125</volume>
          /paper 82.pdf
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Hochreiter</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schmidhuber</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>Long short-term memory</article-title>
          .
          <source>Neural computation 9(8)</source>
          ,
          <volume>1735</volume>
          {
          <fpage>1780</fpage>
          (
          <year>1997</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Ionescu</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          , Muller, H.,
          <string-name>
            <surname>Peteri</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cid</surname>
            ,
            <given-names>Y.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liauchuk</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kovalev</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Klimuk</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tarasau</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Abacha</surname>
            ,
            <given-names>A.B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hasan</surname>
            ,
            <given-names>S.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Datla</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Demner-Fushman</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dang-Nguyen</surname>
            ,
            <given-names>D.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Piras</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Riegler</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tran</surname>
            ,
            <given-names>M.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lux</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gurrin</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pelka</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Friedrich</surname>
            ,
            <given-names>C.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>de Herrera</surname>
            ,
            <given-names>A.G.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garcia</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kavallieratou</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>del Blanco</surname>
            ,
            <given-names>C.R.</given-names>
          </string-name>
          , Rodr guez, C.C.,
          <string-name>
            <surname>Vasillopoulos</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Karampidis</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chamberlain</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Clark</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Campello</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>ImageCLEF 2019: Multimedia retrieval in medicine, lifelogging, security and nature</article-title>
          . In:
          <article-title>Experimental IR Meets Multilinguality, Multimodality, and Interaction</article-title>
          .
          <source>Proceedings of the 10th International Conference of the CLEF Association (CLEF</source>
          <year>2019</year>
          ),
          <source>LNCS Lecture Notes in Computer Science</source>
          , Springer, Lugano,
          <source>Switzerland (September 9-12</source>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Pedregosa</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Varoquaux</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gramfort</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Michel</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Thirion</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grisel</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blondel</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Prettenhofer</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weiss</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dubourg</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vanderplas</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Passos</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cournapeau</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brucher</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Perrot</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Duchesnay</surname>
          </string-name>
          , E.:
          <article-title>Scikit-learn: Machine learning in Python</article-title>
          .
          <source>Journal of Machine Learning Research</source>
          <volume>12</volume>
          ,
          <volume>2825</volume>
          {
          <fpage>2830</fpage>
          (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Sun</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chong</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tan</surname>
            ,
            <given-names>Y.X.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Binder</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Imageclef 2017: Imageclef tuberculosis task - the sgeast submission</article-title>
          .
          <source>In: Working Notes of CLEF 2017 - Conference and Labs of the Evaluation Forum</source>
          , Dublin, Ireland,
          <source>September 11-14</source>
          ,
          <year>2017</year>
          . (
          <year>2017</year>
          ), http://ceur-ws.
          <source>org/</source>
          Vol-1866/paper 130.pdf
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>