<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Cataplexy Detection: Neurologists, You Are Not Alone!</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>(Discussion Paper)</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ilaria Bartolini</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andrea Di Luzio</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science and Engineering (DISI), Alma Mater Studiorum, University of Bologna</institution>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2021</year>
      </pub-date>
      <fpage>5</fpage>
      <lpage>9</lpage>
      <abstract>
        <p>Narcolepsy with cataplexy is a severe lifelong disorder characterized, among the others, by the sudden loss of bilateral face muscle tone triggered by emotions (cataplexy). In this extended abstract, we present two methodologies for the automatic analysis of patients' videos able to assist neurologists in diagnosing the disease and/or detecting attacks. Indeed, recent findings demonstrated that the detection of abnormal motor behaviors in video recordings of patients undergoing emotional stimulation is efective in characterizing the disease symptoms. Such motor behaviors (ptosis, mouth opening, head drop) are however to be discovered by neurologists through manual inspection of patients' videos. Automatic content-based video analysis is clearly of immediate help here. Experimental results conducted on real data support the efectiveness of the presented automated techniques.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Video-based classification of cataplexy</kwd>
        <kwd>automatic video content analysis</kwd>
        <kwd>motor behavior patterns</kwd>
        <kwd>data analysis for health</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Narcolepsy with cataplexy is a rare disorder mainly arising in young adults/children
characterized by daytime sleepiness, sudden loss of muscle tone while awake triggered by emotional
stimuli (cataplexy), hallucinations, sleep paralysis, and disturbed nocturnal sleep [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. A recent
approach for the detection of the disease is based on an analysis of video recordings of patients
undergoing emotional stimulation made on-site by medical specialists [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. According to this
methodology, cataplexy is present if any of three abnormal motor behaviors is detected in the
patient video: ptosis (a drooping or falling of the upper eyelid), head drop, and smile/mouth
opening [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. Such patterns are, however, still to be manually detected by neurologists through
visual inspection of videos. This is due to the complete absence of automatic technological
solutions able to properly support neurologists in such a delicate task.
      </p>
      <p>
        It is evident that a tool able to detect the “correct” facial expression changes (i.e., the disease
symptoms) from video recordings of patients would be able to automatically identify the
presence of the disease. This could be extremely helpful, not only to support neurologists in
diagnosing the disease, but also in monitoring everyday activities in a non-invasive way to
provide early warnings in the event of the insurgence of a crisis. Indeed, it is well known that
the synergistic use of Machine Learning (ML) techniques can help in alleviating the burden
of the medical specialist in analyzing patient data, thus improving diagnostic consistency and
accuracy [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], we introduced the CAT-CAD (Computer Aided Diagnosis for CATaplexy) tool, which
exploits ML techniques for the automatic analysis of video recordings made on patients
undergoing emotional stimulation through the vision of funny movies designed to evoke the
laughter. By means of a user friendly GUI, CAT-CAD efectively supports neurologists with (1)
the automatic detection of disease symptoms, and thus in the disease recognition/monitoring,
and (2) advanced functionalities for video playback and browsing/retrieval. CAT-CAD is the
ifrst tool to allow the automatic recognition of cataplexy symptoms based on the analysis of
patients’ video recordings.
      </p>
      <p>
        In this extended abstract, we report details on the video analyzer for the automatic detection
of cataplexy symptoms. This component of the CAT-CAD system is built on top of SHIATSU, a
general and extensible framework for video retrieval which is based on the (semi-)automatic
hierarchical semantic annotation of videos exploiting the analysis of their visual content [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ],
and exploits features managed through the Windsurf software library [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>After reviewing some background information, we detail the methodologies used to
automatically analyze videos: a pattern-based technique able to recognize facial patterns and
a novel approach based on convolutional neural networks (Section 2). Finally, we provide
results obtained from an extensive experimental evaluation to compare the performance of the
two video analysis approaches, using a benchmark containing recordings from real patients
(Section 3) and conclude (Section 4).</p>
      <sec id="sec-1-1">
        <title>1.1. Background and Related Work</title>
        <p>
          Narcolepsy with cataplexy usually arises in adolescence or young adulthood, but the diagnosis
is typically established after a long period with a mean delay (across Europe) from symptom
onset to diagnosis of 14 years [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ]. The diagnosis delay is due not only to the failure to recognize
the symptoms of the disease, but also to the misinterpretation of cataplexy phenomena as the
expression of other disorders, such as episodes of loss of consciousness of epileptic nature, force
reductions due to neuromuscular disorder or behavioral disorders of childhood psychiatric or
neuropsychiatric relevance.
        </p>
        <p>
          Few scientific studies have considered the video-polygraphic features of cataplexy in adult
age and only recently the motor phenotype of childhood cataplexy has been described
exclusively using video recordings of the attacks evoked by watching funny cartoons [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. These
studies showed that in the context of the physiological response to the laughter there are the
distinctive elements of cataplexy, called motor behaviors patterns, particularly evident at the
level of the facial expression changes. In particular, the three most recurrent motor phenomena
(often displayed by patients afected by the disease) are ptosis, head drop, and smile/mouth
opening [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ].
        </p>
        <p>
          To the best of our knowledge, CAT-CAD is the first study about the automatic recognition
of cataplexy by exploiting patient video recordings. However, automatic detection of facial
motor phenomena similar to the ones used to diagnose cataplexy is commonly used in other
contexts. For example, the detection of eyelid closure, head pose, or mouth opening is useful
for the automatic recognition of fatigue/drowsiness in vehicle drivers [
          <xref ref-type="bibr" rid="ref11 ref6 ref9">11, 9, 6</xref>
          ]. The “verbatim”
use of such techniques in the context of cataplexy diagnosis is however inappropriate, since
the peculiar motor patterns are somewhat diferent, even if they can be detected using similar
facial features.
        </p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>2. The CAT-CAD Video Analyzer</title>
      <p>The core of the CAT-CAD tool is the real-time analysis of patients’ videos to detect the presence
of disease symptoms (i.e., ptosis, head drop, and smile/mouth opening). Two diferent approaches
were developed to perform video analysis, with the idea of comparing their relative performance
and possibly combining them to achieve the best possible result in recognizing the diferent
motor phenomena:
• The Pattern-Based approach (Section 2.1) is built on the automatic detection of facial
features in video frames.
• The Deep Learning approach (Section 2.2) exploits three convolutional neural networks,
each trained to detect a specific motor phenomenon.</p>
      <sec id="sec-2-1">
        <title>2.1. Video Analyzer: Pattern-Based Approach</title>
        <p>
          The first video analyzer to be implemented in CAT-CAD for the detection of cataplexy motor
phenomena exploits facial landmarks, as detected by OpenFace [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. The first step of the pattern
characterization process consists in detecting and extracting patients’ facial landmarks of
interest from each video frame; this is necessary because it is safe to assume that diferent
patients have diferent facial features. Using OpenFace on each single video frame we are thus
able to extract, for each video, a time series of multi-dimensional feature vectors, each vector
characterizing the facial landmarks extracted from a single frame.
2.1.1. Ptosis
Ptosis is a drooping or falling of the upper eyelid. This, however, should not be mistaken
as a (regular) eye blink. For this, ptosis is detected whenever eyes are closed for a period of
time longer than a typical blink. For each frame, a 12-dimensional feature vector is extracted,
containing the (, ) coordinates of six landmarks characterizing the shape of the eye. The Eye
Aspect Ratio (EAR) can then be defined as the ratio of the eye height to the eye width (averaged
for left and right eye). EAR is partially invariant to head pose and fully invariant to uniform
image scaling and in-place face rotation. The semantics of EAR are as follows: when an eye
is closing, EAR approaches zero, whereas when the eye is completely open, EAR attains its
maximum value (which varies from person to person). Therefore, we define the presence or
absence of ptosis by measuring the length of the time series corresponding to a “long enough”
sequence of frames with closed eyes (EAR lower than a threshold).
2.1.2. Head Drop
For head drop, a 8- feature vector is extracted from each frame, including the (, )
coordinates of the two landmarks characterizing the external corner of each eye and the one that is
immediately below the tip of the nose and the rotation of the head around  and  axes. The
Center of Gravity (CoG) of the three landmarks is then used to measure rotation around the 
axis. Head drop is then detected if rotation around one of the three axes exceeds a threshold.
2.1.3. Smile/Mouth Opening
For the third motor phenomenon, a 8- feature vector is extracted for each frame, containing
(, ) coordinates of four landmarks characterizing the shape of the mouth. The Mouth Aspect
Ratio (MAR) can then be defined as the ratio of the mouth width to the mouth height. Like
EAR, MAR is partially invariant to head pose and fully invariant to uniform image scaling and
in-place face rotation. When the mouth is closed, MAR attains its maximum value (which varies
from person to person), while if the mouth is completely open, MAR reaches its lowest value;
intermediate values characterize various types of smile. We thus consider the cataplectic mouth
opening as present if the current MAR is lower than a threshold, indicating that the patient is
smiling widely or opening her mouth.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Video Analyzer: Deep Learning Approach</title>
        <p>
          The alternative video analysis tool is based on convolutional neural networks (CNNs). The
CNN architecture used in this work is based on the DeXpression Network [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ], which achieves
excellent performance in expression recognition, and has been implemented using TensorFlow
(https://www.tensorflow.org/).
        </p>
        <p>
          Our CNN architecture consists of three diferent types of blocks:
1. an Input Block, which performs image pre-processing,
2. a Feature Extraction Block, inspired by the architectural principles introduced by GoogleNet [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ],
which is repeated four times, and
3. an Output Block, which is used to produce the result class from the features extracted by
previous layers.
        </p>
        <p>We have trained three diferent networks, one for each motor phenomenon to be recognized.
The three networks share the same architecture, but the learned weights are clearly diferent,
due to the use of diferent training classes. It is clear that, for this approach, each frame is
analyzed per se, and no considerations on frame sequences, like duration of eyelids closing or
of head drop, can be extrapolated, contrary to the pattern-based approach. The three neural
networks have been trained for 8 epochs each, for a total time of about 12 hours.</p>
        <p>For each video frame, the face of the patient is first detected by means of OpenFace and then
cropped. The resulting image is converted to grayscale and downsized to produce images of
320 × 320 pixels. Cropping of the images was necessary in order to provide the CNN with only
face details (thus avoiding that the surrounding environment would distract the learning).</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Experimental Evaluation</title>
      <p>
        The benchmark used for our experimental evaluation consists of a population of patients of
the Outpatient Clinic for Narcolepsy of the University of Bologna, who were assessed for the
presence of cataplexy by way of a neurophysiological and biological diagnosis [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ].
      </p>
      <p>The first (experimental) group of patients includes 14 subjects displaying symptoms of the
disease. Training of video analyzers has been performed using an inter-patient separation
scheme, where patients have been randomly assigned to non-overlapping training and test
sets, by respecting sex and age distribution. In particular, 11 patients have been included in
the training set (thus, their entire videos have been used to train each analyzer), while the
remaining 3 patients have been exploited to test the accuracy of the tool.</p>
      <p>The second group includes 44 diferent subjects that show no sign of the disease. Among
those, 14 patients have been selected as a control group so as to follow the same sex and age
distribution of the experimental group.</p>
      <p>For the deep-learning approach, data augmentation was performed by adding, to each training
set frame, seven additional images by performing: (i) 3 rotations with a random angle between
− 45∘ and +45∘ , (ii) 3 translations with a random shift between -50 and 50 pixels, and (iii) 1
horizontal flipping . The final training sets consists of 191140 labeled images for ptosis, 61216
labeled images for head-drop, and 108196 labeled images for mouth opening.</p>
      <sec id="sec-3-1">
        <title>3.1. Performance Measures</title>
        <p>To objectively evaluate the performance of our analyzers, each frame can be labeled according to
a confusion matrix as correctly and incorrectly recognized for each of the two classes available
(in our case, motor phenomenon actually present or absent). The four possible outcomes
are  (true positive, a frame where the symptom is correctly detected as present),   (false
negative, symptom wrongly not detected),  (true negative, symptom correctly not detected)
and   (false positive, symptom wrongly detected as present). From the confusion matrix, the
performance measures used in our experiments are defined as follows.</p>
        <p>Recall/Sensitivity () is defined as the fraction of the frames showing signs of the disease

(positives) that are correctly identified:  = + .  is therefore used to measure the
accuracy of a technique in recognizing the presence of the disease.</p>
        <p>Specificity () is the fraction of frames not showing the disease (negatives) that are correctly

classified:  = + .  thus expresses the ability of a technique to avoid false alarms
(which can lead to expensive/invasive exams).</p>
        <p>Precision ( ) is another popular metric, besides  and , which are the fundamental
prevalenceindependent statistics.  is defined as the fraction of correct positively classified frames
and assesses the predictive power of the classifier:  = + .</p>
        <p>Accuracy () measures the fraction of correct decisions, to assess the overall efectiveness of
+
the algorithm:  = +++ .</p>
        <p>Balanced Score (1) is a commonly used measure, combining  and  in a single metric
1 2· 
computed as their harmonic mean: 1 = 2 1 + 1 = 2· ++ .</p>
        <p>
          For the pattern-based approach, thresholds used for the detection of ptosis, head drop, and
mouth opening were chosen as the ones providing the best classifying performance on the test
set [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ]. To this end, a Receiver Operating Characteristics (ROC) graph is used for each threshold,
and the threshold value maximizing the harmonic mean of  and  measures is chosen as
the optimal one: this represents a metric for imbalanced classification, seeking an equilibrium
between the two measures [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ].
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Overall Performance</title>
        <p>• The pattern-based approach lead to significantly superior results with respect to its deep
learning counterpart. In particular, for cataplectic subjects the former attains the best
performance in 85% of the metrics (17 out of 4 × 5 = 20 performance measures).
• When considering specific motor phenomena, the pattern-based approach consistently
outperforms the deep learning approach in detecting ptosis, while the latter sports superior
measures only for specificity and precision in detecting mouth opening and for recall in
head drop detection.
• The superior specificity of the pattern-based technique is confirmed in non-cataplectic
subjects, with an overall specificity at 98%.</p>
        <p>A possible explanation for the inferior performance of the deep learning approach is the fact
that such approach cannot discriminate between quick and long eye blinks/head drops, due
to the fact that each frame is analyzed individually by the CNN. It is therefore likely that the
higher number of false positives is because the CNN wrongly detects “regular” eye blinks or
head movements as ptosis or head drop.</p>
        <p>For the case of non-cataplectic subjects, it is interesting to note that the performance of
the deep learning approach for the overall detection of cataplexy is sensibly worse than those
attained for the single motor phenomena. For such patients, false positives for ptosis, head drop,
and mouth opening are present in diferent frames. Indeed, due to the absence of positive cases,
the set of false positive frames for the overall cataplexy coincides with the union of frames
wrongly classified by any specific motor phenomenon detector.</p>
        <p>Finally, we include a brief discussion about eficiency of the proposed techniques. On our
experimental setup, which involved a commodity (low-end) machine, we were able to extract
EAR, CoG and MAR descriptors in real-time for each video frame. Clearly, this is the more
time consuming operation for the pattern-based approach, thus it is proven that the whole
process of automatic detection can be performed on-line during a single emotional stimulated
video recording session. On the other hand, our current implementation of the deep learning
approach is only capable to obtain a throughput of 18.5 frame/s, thus being unable to attain
real-time performance (recall that the frame rate of videos is 30 frame/s). The reason for this
measure is the following: when analyzing a single frame, about 50% of the time is spent in
detecting the position of the patient face, about 25% in cropping the image (retaining only the
face), and 25% for the classification of the frame by the three neural networks. The bottleneck
of the whole computation is clearly the face detection phase, which we implemented using the
OpenFace library, instead of using other faster methods (such as the well-known Haar-Cascade
iflter). This choice was carried out starting from the consideration that quicker filters often
fail to identify the face within the image, especially in videos with excessive head movement,
which is the common case for cataplectic subjects.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusions</title>
      <p>In this extended abstract, we reported details on the video analyzer of CAT-CAD for the
automatic detection of cataplexy symptoms (ptosis, head drop, and smile/mouth opening). Two
diferent approaches are introduced for the detection of disease symptoms: the Pattern-Based
approach is based on analysis of facial features, using the OpenFace framework, while the
Deep Learning approach uses CNNs, as implemented by TensorFlow. An extensive comparative
experimental evaluation conducted on a benchmark of real patients recordings demonstrated
the accuracy of the proposed techniques. When comparing the efectiveness of the two video
analyzers we introduced to detect cataplexy symptoms, the pattern-based approach achieves
superior performance. One of the possible explanations for the inferior detection accuracy of
the deep learning approach is the fact that 2D CNNs are unable to properly take into account
the temporal dimension that correlates subsequent frames in a video. The use of 3D CNNs could
be an interesting way to pursue, and we plan to consider their inclusion in CAT-CAD.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>T.</given-names>
            <surname>Baltrušaitis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Zadeh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. Yao</given-names>
            <surname>Chong</surname>
          </string-name>
          , L.-P.; Morency, OpenFace
          <volume>2</volume>
          .0:
          <string-name>
            <given-names>Facial</given-names>
            <surname>Behavior Analysis Toolkit</surname>
          </string-name>
          ,
          <source>in: Proceedings of FG</source>
          <year>2018</year>
          ,
          <article-title>Xi'an, China</article-title>
          , May
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>I.</given-names>
            <surname>Bartolini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Di</surname>
          </string-name>
          <string-name>
            <surname>Luzio</surname>
          </string-name>
          ,
          <article-title>CAT-CAD: A Computer-Aided Diagnosis Tool for Cataplexy</article-title>
          , Computers,
          <year>2021</year>
          ,
          <volume>10</volume>
          (
          <issue>4</issue>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>I.</given-names>
            <surname>Bartolini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Patella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Romani</surname>
          </string-name>
          , SHIATSU: Tagging and Retrieving Videos Without Worries,
          <source>Multimedia Tools and Applications</source>
          ,
          <year>2013</year>
          ,
          <volume>63</volume>
          (
          <issue>2</issue>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>I.</given-names>
            <surname>Bartolini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Patella</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Stromei, The Windsurf Library for the Eficient Retrieval of Multimedia Hierarchical Data</article-title>
          ,
          <source>in: Proceedings of SIGMAP</source>
          <year>2011</year>
          , Seville, Spain,
          <year>July 2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>I. Bartolini</surname>
          </string-name>
          et al,
          <source>Automatic Detection of Cataplexy, Sleep Medicine</source>
          ,
          <year>2018</year>
          ,
          <volume>52</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>L.M.</given-names>
            <surname>Bergasa</surname>
          </string-name>
          et al,
          <article-title>Analysing Driver's Attention Level Using Computer Vision</article-title>
          , in
          <source>: Proceedings of ITSC</source>
          <year>2008</year>
          , Beijing, China,
          <year>June 2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>P.</given-names>
            <surname>Burkert</surname>
          </string-name>
          et al,
          <source>DeXpression: Deep Convolutional Neural Network for Expression Recognition</source>
          , arXiv,
          <year>September 2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>T.</given-names>
            <surname>Fawcett</surname>
          </string-name>
          ,
          <article-title>An Introduction to ROC Analysis</article-title>
          ,
          <source>Pattern Recognition Letters</source>
          ,
          <year>2006</year>
          ,
          <volume>27</volume>
          (
          <issue>8</issue>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J.</given-names>
            <surname>Jo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.G.</given-names>
            <surname>Jung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ryoung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <article-title>Vision-Based Method for Detecting Driver Drowsiness and Distraction in Driver Monitoring System</article-title>
          , Optical Engineering,
          <year>2011</year>
          ,
          <volume>50</volume>
          (
          <issue>12</issue>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>L.</given-names>
            <surname>Lazli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Boukadoum</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.A.</given-names>
            <surname>Mohamed</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          <article-title>Survey on Computer-Aided Diagnosis of Brain Disorders through MRI Based on Machine Learning and Data Mining Methodologies with an Emphasis on Alzheimer Disease Diagnosis and the Contribution of the Multimodal Fusion</article-title>
          , Applied Sciences,
          <year>2020</year>
          ,
          <volume>10</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>B.</given-names>
            <surname>Mandal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <source>Towards Detection of Bus Driver Fatigue Based on Robust Visual Analysis of Eye State, IEEE Transactions on Intelligent Transportation Systems</source>
          ,
          <year>2017</year>
          ,
          <volume>18</volume>
          (
          <issue>3</issue>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>F.</given-names>
            <surname>Pizza</surname>
          </string-name>
          et al,
          <article-title>Clinical and Polysomnographic Course of Childhood Narcolepsy with Cataplexy</article-title>
          ,
          <source>Brain</source>
          ,
          <year>2013</year>
          ,
          <volume>136</volume>
          (
          <issue>12</issue>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>G.</given-names>
            <surname>Plazzi</surname>
          </string-name>
          et al,
          <source>Complex Movement Disorders at Disease Onset in Childhood Narcolepsy with Cataplexy</source>
          ,
          <source>Brain: A Journal of Neurology</source>
          ,
          <year>2011</year>
          ,
          <volume>134</volume>
          (
          <issue>12</issue>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>F.</given-names>
            <surname>Provost</surname>
          </string-name>
          ,
          <source>Machine Learning from Imbalanced Data Sets 101, in: Proceedings of the AAAI'2000 Workshop on Imbalanced Data Sets</source>
          ,
          <volume>68</volume>
          (
          <year>2000</year>
          ), Austin, TX,
          <year>July 2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>C.</given-names>
            <surname>Szegedy</surname>
          </string-name>
          et al,
          <article-title>Going Deeper with Convolutions</article-title>
          ,
          <source>in: Proceedings of CVPR</source>
          <year>2015</year>
          , Boston, MA, June
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>S.</given-names>
            <surname>Vandi</surname>
          </string-name>
          et al, A Standardized Test to Document Cataplexy,
          <source>Sleep Medicine</source>
          ,
          <year>2019</year>
          ,
          <volume>53</volume>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>