<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>ImageCLEF 2019: A 2D Convolutional Neural Network Approach for Severity Scoring of Lung Tuberculosis using CT Images</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of CSE, SSN College of Engineering</institution>
          ,
          <addr-line>Kalavakkam-603110</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Tuberculosis (TB) is an air-borne disease, which affects the lungs and often spreads through sputum. According to the report of World Health Organization 9 million people world-wide are affected with TB. Tuberculosis can be cured easily when diagnosed in its early stage and with accurate CT Analysis. As an effort to form a technical forum for effective analysis and diagnosis, ImageCLEF released the Tuberculosis 2019 tasks, each dealing with one aspect of understanding and tackling the disease. We have taken up one sub-task that aims at assessing the severity of the tuberculosis disease as low or high. The task is implemented using a deep neural network approach using 2-D Convolutional Neural Network (CNN) with appropriate preprocessing. The CT volumes are segmented with the provided masks and further pre-processed with the aid of med2image, a python utility to obtain slices of CT scans, prior to training the model. The best run of the proposed CNN model resulted with an accuracy of 0.607 and an AUC of 0.626. The achieved result is placed 9th in the overall leaderboard of the ImageCLEF 2019 Tuberculosis challenge for severity scoring .</p>
      </abstract>
      <kwd-group>
        <kwd>Severity scoring</kwd>
        <kwd>Lung tuberculosis</kwd>
        <kwd>Pre-processing</kwd>
        <kwd>Lung-mask</kwd>
        <kwd>CNN</kwd>
        <kwd>AUC</kwd>
        <kwd>Accuracy</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Tuberculosis (TB) is an airborne disease that affects the lungs. Often spread through
sputum, cough and infected droplets, it is quite widespread affecting about 9 million
world-wide. The treatment depends upon the degree of infection, i.e the severity [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
The severity evaluation has been executed by medical practitioners via a diverse set
of devices including mycobacterial culture test, pleural fluid and cerebrospinal fluid
analysis, lesion patterns obtained from radiological images of lungs besides
individualistic factors such as the patient’s age, prior treatment etc. Computed Tomography
(CT) is widely used for analysis of the lesion patterns. Besides being prone to errors,
a manual approach can prove to be costly, both in terms of capital and time. A
computerized method, on the other hand upholds time efficiency and precision. In this
paper, a Convolutional Neural Network (CNN) approach for severity scoring of lung
tuberculosis based on CT scans is discussed with results. This work is a subtask of
the tuberculosis tasks of ImageCLEF 2019 [
        <xref ref-type="bibr" rid="ref2 ref8">2, 8</xref>
        ]. This work establishes a standard
scale against which evaluation of the CT in subject can be done for determining the
severity.
      </p>
      <p>The sections span across following: Section 1 gives a brief introduction about the
importance of this problem and the necessity to find the severity of tuberculosis.
Section 2 gives a glimpse of the dataset and how it is spread across the two classes
and Section 2.1 details about the data preprocessing procedures. Section 3 explains
the proposed model using convolutional neural network with the parameters chosen
for analysis. In Section 4, the results of various runs are discussed. Finally, Section 5
concludes this paperwork and looks into the futuristic aspects for further
improvisation of the proposed model.</p>
    </sec>
    <sec id="sec-2">
      <title>2 Dataset</title>
      <p>
        In this edition of ImageCLEF 2019 TB tasks, the dataset contains 335 chest CT
scans of TB patients along with a set of clinically relevant metadata, where data of
218 patients are used for training and 117 for testing. For all patients, 3D CT images
are stored in the compressed NIFTI (Neuroimaging Informatics Technology
Initiative) file format with a slice size of 512×512 pixels and the number of slices varies
from 50 to 400 for each patient. This file format stores raw voxel intensities in
Hounsfield Units (HU) as well the corresponding image metadata like image
dimensions, voxel size in physical units, slice thickness, etc. The selected metadata
includes the following binary measures: disability, relapse, symptoms of TB,
comorbidity, bacillary, drug resistance, higher education, ex-prisoner, alcoholic, smoking
and severity score ranges from 1 to 5 assigned by medical doctors. To treat this task
as a binary classification problem, the severity scores are grouped as high severity
with scores 1, 2 and 3, and low severity with scores 4 and 5. Moreover, for all
patients automatic extracted masks of the lungs are provided. In Table 1, the number of
patients of each severity class in the training set and the number of patients in test
set is given [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
2.1 Data Preprocessing
The dataset for the TB tasks are given in compressed NIfTI (Neuroimaging
Informatics Technology Initiative) format. Initially, the file is decompressed and the
slices were extracted using med2image, a Python utility. For each Nifti image we obtain
a certain number of slices ranging from 50 to 400 jpeg images. The lung masks
provided by the organizers are used, to avoid potential confusion resulting from
identification of similar structures resembling lungs in other parts of CT images. The next
step involves masking the images. The given masks are converted to grayscale
format and each pixel is checked individually; if the pixel is not black it is converted to
white. In this way, a final mask is created, with pixels of two values such as black
(0) or white (255). Now the original scan of the lung is converted to grayscale and
each pixel of it is multiplied with the corresponding pixel in the created final mask
using bitwise and operation. Thus, the lungs are segmented from the original scans
[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. On the other hand, not all slices necessarily contain relevant information that
can be useful to identify severity of TB. For the same reason, it is essential to filter
slices to preserve only those that can be informative and contain relevant
information. Upon visual inspection, slices ranging between 55 and 85 are used and other
slices were eliminated from further processing. The slices being ordered, the 31
most informative usually fall at the center of the list. The workflow of the
preprocessing stages is given in Figure 3.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3 Methodology</title>
      <p>
        Convolutional Neural network takes an image as input, passes it through a series of
convolutional layers, nonlinear activation layers, pooling (downsampling) and a
fully connected layer to output the classification labels. It differs from normal neural
network in two aspects: atleast one convolutional layer and filters. The model for TB
severity scoring is created using 2-D convolutional neural network using software
libraries Keras [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] with Tensorflow [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] for backend. The 3D images of the procured
CT scans are sliced and converted to 2D images in the preprocessing stage. The
network is designed with three 2D convolution layers, rectified linear unit (ReLU)
activation function and each convolution layer followed the max pooling layer.
These led to a complete layer structure which is connected to 1000 outputs with weight
by the dense layer with ReLU activation. Finally, these activations run through a
softmax layer, which output a tensor of size 2, for each category. Binary
crossentropy is used as loss function with Adam and RMSProp as optimizers. The model
files are built for different runs by varying the hyper parameters of the base model.
The corresponding CNN design structure is shown in Figure 4. In each layer the
values are mentioned from the model summary of one run for more clarity.
      </p>
    </sec>
    <sec id="sec-4">
      <title>4 Experiments and Results</title>
      <p>The CNN model is trained by varying the hyperparameters such as number of filters,
epochs and optimizers. The runs had a filter size of 64×64 with batch size 32 and
loss type as binary cross entropy. The difference in the accuracy is brought by
changing the epoch value and the optimizers such as Adam and RMSProp. The
different runs of CNN model by varying the hyperparameters is given in Table2.
binary cross
entropy
binary cross
entropy
binary cross
entropy
RMSProp</p>
      <p>Adam</p>
      <p>RMSProp</p>
      <p>RMSProp
Number of epochs
Loss type</p>
      <p>Optimizer</p>
      <p>The intermediate visualization of convoluation layers and max pooling is shown
in Figure 5, for Run1, Patient ID 181 and slice number 65.</p>
      <p>The result of submitted four runs are listed in Table 3, for training, validation and
test dataset with necessary parameters. In testing, 31 slices per patient is considered
as similar to training and validation, for all 117 patients. The probability of high
severity for each patient is calculated from the average of “probability of high” of all
31 slices of the specific patient. For example, the class probability of patient ID 77
in testing, for slice number 60 is represented as [0.1.]. Here, “0.” is the probability of
having low severity and “1.” is the probability of having high severity. We find the
probability of having high severity for each of 31 slices of the patient and then
computed the average of it. The average probability of high severity for patient ID 77 in
each run is given in Table 4.</p>
      <p>The test dataset results are evaluated using two metrics namely accuracy and Area
Under ROC Curve (AUC) and ranking is carried out among the participated teams.</p>
      <p>For better visualization, the same information is plotted and shown as graphs in
Figures 6 and 7 using Tableau Tool. In Figure 6, the value of evaluation metrics for
test set is given for all four runs. From the graph, it is clearly visible that run 3 has
higher AUC and accuracy than remaining runs. In Figure 6, the accuracy for
training, validation and testing dataset are given for all four runs. From the graph, it is
clearly visible that run 3 has higher value in all the cases. In addition, run 4 has
higher training and validation accuracy, but the test accuracy is low than run 2,
might have occurred due to the chosen filter size of each layer.</p>
    </sec>
    <sec id="sec-5">
      <title>5 Conclusion and Future Work</title>
      <p>In this paper, analysis of severity scoring (SVR) subtask for lung tuberculosis using
2D Convolutional Neural Network is implemented. The classification results
obtained for the given set 3D CT Images are submitted for evaluation. In our approach,
preprocessing of the dataset has been carried out to convert the images into 2D
slices, and the images are split into training and validation set. The proposed model is
built using CNN, trained and validated using tuning the hyperparameters for four
different runs. From the runs submitted, the primary run is ranked 9th place among
the team participations.</p>
      <p>CNN is a preferred approach, since it facilitates automatic detection of the low
level and high level features, from large training dataset. However, a large dataset
might prove disadvantageous in terms of memory during the training phase. This
can be overcome by the use of a GPU and choosing optimal hyperparameters. In
future, the proposed model can be improvised by considering all the slices of the CT
images, to build the train model using GPU and transfer learning approach.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>1. WHO page, https://www.who.int/news-room/fact-sheets/detail/tuberculosis, last accessed May 2019</mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <article-title>ImageCLEF 2019 page</article-title>
          , https://www.imageclef.org/2019/medical/tuberculosis, last accessed May 2019
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Crowd</surname>
            <given-names>AI</given-names>
          </string-name>
          page, https://www.crowdai.org/challenges/imageclef-2019
          <string-name>
            <surname>-</surname>
          </string-name>
          tuberculosis-severity-scoring/ leaderboards, last accessed May 2019
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>Yashin</given-names>
            <surname>Dicente</surname>
          </string-name>
          <string-name>
            <surname>Cid</surname>
          </string-name>
          ,
          <article-title>Oscar A. Jiménez-del-</article-title>
          <string-name>
            <surname>Toro</surname>
          </string-name>
          ,
          <article-title>Adrien Depeursinge, and Henning Müller, Efficient and fully automatic segmentation of the lungs in CT volumes</article-title>
          . In: Goksel,
          <string-name>
            <surname>O.</surname>
          </string-name>
          , et al. (eds.)
          <source>Proceedings of the VISCERAL Challenge at ISBI. No. 1390 in CEUR Workshop Proceedings (Apr</source>
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>5. Keras documentation, https://keras.io/, last accessed May 2019</mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>6. Tensorflow documentation, https://www.tensorflow.org/, last accessed May 2019</mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>Yashin</given-names>
            <surname>Dicente</surname>
          </string-name>
          <string-name>
            <surname>Cid</surname>
          </string-name>
          , Vitali Liauchuk, Dzmitri Klimuk, Aleh Tarasau, Vassili Kovalev, Henning Müller,
          <source>Overview of ImageCLEFtuberculosis 2019 - Automatic CT-based Report Generation and Tuberculosis Severity Assessment, CLEF 2019 Working Notes. CEUR Workshop Proceedings (CEUR- WS.org)</source>
          ,
          <source>ISSN 1613-0073</source>
          , http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2380</volume>
          /.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>Bogdan</given-names>
            <surname>Ionescu</surname>
          </string-name>
          , Henning Müller, Renaud Péteri, Yashin Dicente Cid, Vitali Liauchuk, Vassili Kovalev, Dzmitri Klimuk, Aleh Tarasau, Asma Ben Abacha, Sadid A.
          <string-name>
            <surname>Hasan</surname>
          </string-name>
          , Vivek Datla, Joey Liu, Dina Demner-Fushman,
          <article-title>Duc-Tien Dang-Nguyen, Luca Piras</article-title>
          , Michael Riegler,
          <string-name>
            <surname>Minh-Triet</surname>
            <given-names>Tran</given-names>
          </string-name>
          , Mathias Lux, Cathal Gurrin, Obioma Pelka,
          <string-name>
            <surname>Christoph M. Friedrich</surname>
          </string-name>
          , Alba García Seco de Herrera, Narciso Garcia, Ergina Kavallieratou,
          <source>Carlos Roberto del Blanco</source>
          , Carlos Cuevas Rodríguez, Nikos Vasillopoulos, Konstantinos Karampidis, Jon Chamberlain, Adrian Clark, Antonio Campello, ImageCLEF 2019:
          <article-title>Multimedia Retrieval in Medicine, Lifelogging, Security and Nature In: Experimental IR Meets Multilinguality, Multimodality, and Interaction</article-title>
          .
          <source>Proceedings of the 10th International Conference of the CLEF Association (CLEF</source>
          <year>2019</year>
          ), Lugano, Switzerland,
          <source>LNCS Lecture Notes in Computer Science</source>
          ,
          <source>Springer (September 9-12</source>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <given-names>Yashin</given-names>
            <surname>Dicente</surname>
          </string-name>
          <string-name>
            <surname>Cid</surname>
          </string-name>
          , Vitali Liauchuk, Vassili Kovalev, Henning Müller, Overview of ImageCLEFtuberculosis 2018 -
          <article-title>Detecting multi-drug resistance, classifying tuberculosis type, and assessing severity score</article-title>
          ,
          <source>CLEF working notes, CEUR</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>