<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Attack originality detector using Machine Learning</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Charlotte Millan</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Lotfi Chaari</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Amel Aissaoui</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Eric Gilbert</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>CS Group - Embedded Industrial cybersecurity systems</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Toulouse INP, IRIT, University of Toulouse</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>Security Information and Event Management (SIEM) systems are critical tools in modern cybersecurity, designed to aggregate, analyze, and correlate security data for real-time threat detection. While efective against known and over threats, SIEM systems often struggle to detect subtle variations in attack patterns. These variations, though minor, can signal the evolution of an attack or serve as precursors to more severe attacks, such as advanced persistent threats (APTs). Addressing this gap requires innovative detection methods capable of identifying such "original" attacks. In this paper, we propose a model combining multiple weak classifiers, each employing a diferent classification technique to highlight features indicative of original attacks. A voting system is then applied to the results, classifying as "original" attack segments that fail to achieve consensus. Our method was tested and validated using DoS and Probes attack groups in the NSL-KDD dataset, a benchmark for intrusion detection systems.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Artificial Intelligence</kwd>
        <kwd>Machine Learning</kwd>
        <kwd>Cybersecurity</kwd>
        <kwd>classification</kwd>
        <kwd>Voting system</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Today, malicious actors only need a functioning attack scenario to achieve their objectives, and the
more interconnected objects are, the greater the attack surface (number of exploitable entry points
and vulnerability sources). To cope with the ever-increasing number and complexity of attacks on
their information systems, business and public authorities are turning to platforms for monitoring
their network operations: Security Information and Event Management (SIEM) software solutions.
They must be capable of continuously collecting and aggregating large quantities of data indicating
possible security flaws, known as alerts. Captured at diferent points on the network and emitted by
heterogeneous sources (antivirus, firewalls, network probes, authentication servers, etc.), these alerts
have three main characteristics: velocity, volume, and variety.</p>
      <p>
        One of the essential functions of SIEMs is to correlate the security alerts generated by the same
intrusion attempt, by consolidating alerts and contextualizing them. Clear information can therefore be
provided to Security Operation Center (SOC) analysts who are responsible for decision making and
often overwhelmed by the exploitation of 0-days vulnerabilities and the number of false positives. The
use of data science in cybersecurity can help to correlate events, identify recurring patterns, and detect
abnormal behaviors to improve SIEM performances [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], the authors propose a systematic approach to convert scores from models into probabilities ready
for use in cybersecurity models. To do this, they consider each decision model as a classifier. Thanks to
their multiple-score calibration approach, the scores from their three systems can be combined and
converted into a single probability measure that is more relevant for decision-making. They conclude
that Platt’s logistic regression performs acceptably over a wide range of experiments. They also point
out that multi-score calibration and score extension to higher dimensions, both increase performance
in most cases.
      </p>
      <p>
        The study in [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] uses anomaly scores to detect Advanced Persistent Threats (APTs). To predict the attack
stage, the authors proposed an approach for detecting APT occurrences and systematically mapping the
corresponding events to stages in the execution chain, using anomaly scoring and Machine Learning
(ML) based on a Bayesian network model. Although his method works with raw data, it is more efective
when integrated with SIEM. To cope with Distributed Denial of Services (DDoS) attacks, the authors in
[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] propose a method using various supervised ML methods for classification and prediction. Using
Random Forest models with an accuracy of 99 %, a -Nearest-Neighbors (KNN) method with 96 % and
a Logistic Regression method with 85 %, they obtained good results for DoS attack detection where
experiments have been conducted on the NSL-KDD dataset. As regards unsupervised learning, the
authors in [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] emphasis on the fact that such methods learn what is normal for a dataset and are
therefore able to find variations in unclassified sets such yet unknown attack. To detect anomalies, they
use a combination of six unsupervised ML models. They conclude that unsupervised ML methods are
eficient for generalizing and detecting unknown patterns but may fail to control false positive rates.
The authors therefore recommend using hybrid models that combine supervised and unsupervised
learning.
      </p>
      <p>Based on the current state of the art, it becomes clear that existing methods mainly target well-known
attacks, leaving a critical gap in detecting more nuanced, evolving threats. SIEM systems often struggle
to identify subtle variations in attack patterns that may indicate the evolution of an attack or act as
early warnings for more severe threats. To address this limitation, our work focuses on detecting
these variations, characterizing what we define as "original attacks". We believe that detecting these
original attacks will enhance threat detection capabilities by uncovering previously unnoticed attack.
Accordingly, we propose a novel method to classify attacks originality using a hierarchical classification
model made up of three weak classifiers and a voting system to point out consensual classifications of
attacks based on recorded features.</p>
      <p>In this work, we propose a method to classify attacks originality using a hierarchical classification
model made up of three weak classifiers and a voting system to point out consensual classifications of
attacks based on recorded features. We considered that’s original attack will be a sign of evolution of a
well-known attack, or the first of an advanced attack, and detecting them will be improved the SIEM.
We start working on this subject to find low signals, premises of APT attacks. This paper is organized
as follows. We first present the used data and the proposed method in Section 2. Experimental results is
then presented in Section 3. Finally, conclusions and future work are drawn in Section 4.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Materials and Methods</title>
      <p>Detecting attack originality is a complex challenge, as originality cannot be easily defined or analytically
modeled. To address this, we adopt a data-driven approach, leveraging ML method, to learn from data
in an unsupervised learning framework. We first describe the used NSL-KDD dataset, and then go
through the proposed model.</p>
      <sec id="sec-2-1">
        <title>2.1. Materials</title>
        <p>
          The data used in this study are extracted from the NSL-KDD database [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]. Each entry in this dataset is
made up of 41 features, 38 of which are in numeric format. The other three are relate to the type of
used protocol, the type of connection to the server and the state of the connection at the capture time.
Figures 1 and 2 illustrate two segments from the NSL-KDD dataset, where, for instance, the second and
third features stand for the protocol and service, respectively. The last field, frame in red, indicates the
label associated with each entry: attack type or normal segment.
        </p>
        <p>
          It is worth noting that the NSL-KDD dataset is the most widely used for intrusion detection studies,
as presented by the authors of: [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ],[
          <xref ref-type="bibr" rid="ref7">7</xref>
          ], [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ].
        </p>
        <p>The methodology we have used begins with the presentation of the relevant datasets: in the case of
this study, we need a dataset containing normal attacks and flows, and with the distribution of known
attacks.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Methods</title>
        <p>As illustrated in Figure 2, our method processes attack segments that are precategorized according
to their classes and analyzing them to classify each as either an original or a known (normal) attack.
Figure 3 details the proposed pipeline to analyze data segments and detect attack originality.</p>
        <p>Dataset
attack segment</p>
        <p>Identified attack</p>
        <p>Originality classificator
Well-known attacks</p>
        <p>Originals attacks</p>
        <p>As illustrated in Figure 3, detected attacks are analyzed to decide whether they are original or not,
compared to detected attacks of the same type.</p>
        <p>The proposed methodology is illustrated in Figure 4, where a Principal Component Analysis (PCA)
module is used as a dimension reduction technique in order to project the dimensions of the dataset
into a lower dimension space. Here, we choose two components, as it was enough for our system. The
classification problem is addressed by using  weak classifiers and a voting system. Each classifier 
generates a binary output , defined as follows:
{︃1, if normal</p>
        <p>Each model returns 1 if the segment is normal (usual attack), 0 if is not (original attack), and the
voting system calculates the following weighted decision score:
where  is the number of weak classifiers and ()1≤ ≤  are the weights associated to each weak
classifier. In our case, the same weight is associated to all classifiers ( ∀ ∈ {1, . . . ,  },  = 1/ ).
It is worth noting that a consensus is reached if and only if  ∈ {0, 1}. The decision of the voting
system is set to True (normal attack) if  ∈ {0, 1}. For all the remaining cases, the attack is classified
as original (presents significant deviation from the general behavior of the same class of attack). In our
study, we used three weak classifiers ( 1, 2, 3):
1. K-means (1): classifies data based on their similarity. This unsupervised method does not
consider the type of attack, only the similarity of the segments between each other’s. The number
of clusters has been set to the number of attack types in the dataset. As K-means is not supposed
to detect outliers, we use the distance between centroids of the clusters and the segment itself. A
threshold is adjusted to determine whether a segment may be considered belonging to one of
the clusters or not. Each attack’s type has a maximal distance between the point and the center,
attacks beyond are considered as original.
2. -nearest-neighbors (2) : KNN classifies data according to the distance between neighbors.</p>
        <p>This unsupervised learning algorithm is generally chosen for its simplicity. The number of
neighbors  is set to the number of attack types: to detect original attacks, we took an interest
for the segment misclassified by the algorithm. To define outlier, we use decision boundary:
we define a minimum distance between the points and the boundary, and all points below this
distance are then considered original.
3. One Class SVM (3) : 1C-SVM is an unsupervised classifier that detects outliers using a given
kernel (Gaussian in our case). It returns 0 if the segment is outlier, 1 if it is normal.
For one entry segment, and once the weak classifiers are learned, their outputs ( 1, 2 and 3) are
used as input for the voting system. These inputs are binary integers which take 1 when the attack is
normal, and 0 if original. The voting system classifies a segment as normal attack (non-original) if and
only if all weak classifiers agree.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Experimental results</title>
      <p>
        In this section, we present the experimentation and the results of the proposed method on the widely
used NSL-KDD dataset [
        <xref ref-type="bibr" rid="ref6 ref9">6, 9</xref>
        ].
      </p>
      <sec id="sec-3-1">
        <title>3.1. Protocol</title>
        <p>In the following, we will focus on two groups of attacks: Denial of Services (DoS) and Probes which are
the most presents in the dataset. The other groups of attacks represent 2% of the dataset, which is too
unbalanced to be used without bias. Then, we have divided the dataset in two parts: 80% for training
and 20% for test.</p>
        <p>The dimension reduction step (PCA) has been fit on the training data where diferent numbers of
dimensions have been tested. By cross validation, it turned out that 2 is the most suitable number
of dimensions. The output of the voting system is then analyzed for all segments according to the
evaluation strategy in Section 3.2. Inspections of some non-consensual attacks are also performed by a
cyber analyst in Section 3.2.4. It is worth noting that our system does not focus on attacks classification,
but on detecting their originality. As attacks withing the same class are supposed to be relatively
homogeneous, attacks detected as being original are those which do not match any consensus between
the used classifiers ( 1, 2 and 3).</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Validation of the proposed model</title>
        <p>In the following, results on DoS and Probes attack are detailed. Each type of attack has diferent
sub-types for which originality detection results are provided for weak classifiers and the voting system.
As regards metrics, it is not useful in our case to report precision or F1 score results since our model
does not focus on attack detection, but on their originality. Moreover, no originality annotation is
provided in the used dataset. Annotations only indicate the attack type for each segment. To validate
our results, we report them, then we analyzes them with the manual validation of a data analyst, as
made in the SOC today.</p>
        <sec id="sec-3-2-1">
          <title>3.2.1. Originality detection for Denial of Services attacks</title>
          <p>DoS attacks are the most represented in the NSL-KDD dataset, with 45909 entries divided into five
attack’s type : back, neptune, pod, smurf and teardrop. Table 1 reports the results obtained by the
diferent weak classifiers, as well as the proposed voting system.</p>
          <p>At first sight, we suspected that K-means method might be ineficient, as it detected too few back
attacks. This suspicion was alimented by the distribution of data in each cluster, shown in figure 5.
Since the distribution of segments within clusters did not align with attack types, we initially thought
this mismatch afects the detection of back attacks. But the result obtained" by the anomalies’ detection
models show up a majority of anomalies in back’s attacks, made us conclude the problem is on the
constitution of the attack types, and not in the model.</p>
          <p>We can see in figure 5 that neptune attacks are present in clusters 0, 1 and 4, while teardrop attacks
are in cluster 3 together with smurf and pod ones. Back attacks are in cluster 2. The K-means model
here is eficient for classified attacks that are predominantly present in dataset, in contrast to weakly
represented ones.</p>
          <p>As shown in Table 1, we can see that the voting system works well for neptune attacks, which is
expected because they are massively represented in the database. Back attacks, as are not classified by
K-means, are the most poorly detected by the voting system, as they only achieve consensus in 1% of
cases. Pod and Teardrop, which are better defined, still have a high number of original attacks, with an
average of 50% of segment detected as original attacks.</p>
          <p>t
n
u
co 3000
attack types
neptune
teardrop
back
smurf
pod</p>
          <p>The voting system allows us to take advantage of the three models, and outlines attacks for which no
consensus is reached by the voting system.</p>
        </sec>
        <sec id="sec-3-2-2">
          <title>3.2.2. Originality detection for Probe attacks</title>
          <p>Probe attacks are less numerous than DoS ones, with only four attack’s types (ipsweep, nmap, portsweep
and satan) listed in 11656 attacks. We can see the distribution of the data on Table 2.</p>
          <p>When we analyze results in table 2, K-means model correctly identifies more than 80% of attacks,
and 100% of the ipsweep attacks, based on the number of attacks found by type, on the number of
attacks presents in the dataset. Both models demonstrate good performance, achieving a classification
accuracy, with a good classification rate between 96% and 100%. Around 5% of the attacks demonstrate
the consensus, and the segments show suficient variation to be considered as originals.</p>
        </sec>
        <sec id="sec-3-2-3">
          <title>3.2.3. Comparison with state of the art methods</title>
          <p>
            In order to validate the proposed model, two state of the art methods among those often adopted by
experts are used: Local Outlier factor (LoF) and Isolation Forest (IF). These methods are usually used
as anomaly detectors, and in our case, they are applied to attack data to outline anomalies that could
be considered as a form of originality. These methods are adopted by expert thanks to their eficiency
in anomalies detection, as shown by the authors, in [
            <xref ref-type="bibr" rid="ref10">10</xref>
            ] and [
            <xref ref-type="bibr" rid="ref11">11</xref>
            ], who demonstrate respectively the
eficiency of IF for anomalies detection and cyber-attacks prediction. In [
            <xref ref-type="bibr" rid="ref12">12</xref>
            ] and [
            <xref ref-type="bibr" rid="ref13">13</xref>
            ], authors used LOF
for anomalies detection and network intrusion detection, respectively.
          </p>
          <p>For the DoS attacks, we obtained the distribution shown in table 3</p>
          <p>We see that LoF and IF are not consensual: as LoF seems to be more precise than IF. However, if we
obtained better results than IF for all the attack types, the original attacks detected for back, pod and
teardrop by LoF are less than ours. But our three classifiers share similar results, which confirm the
eficiency of the model. However, back attacks seem to be less prone to original attacks. As our system
detect a majority of original attacks (99%), it is interesting to see that IF models still detects 99,16% of
original attacks, making think that back attacks are dificult to classifies.</p>
          <p>We use the same method for Probes attacks, and the results are presented on table 4</p>
          <p>The state of the art methods seems more eficient for Probes attacks but exception made for nmap
attacks, our methods seem more precise.</p>
          <p>After analysis, we have 80% of segments in common and we choose to examine manually the 20% to
confirm if they are originals, and the results are presented in the section bellow.</p>
        </sec>
        <sec id="sec-3-2-4">
          <title>3.2.4. Validation by a cyber analyst</title>
          <p>We investigate whether attacks classified as original by our model and normal by the competing methods
are indeed original from the expert point of view. We therefore define this three-steps procedure:
1. We first analyze features of normal attacks. Then we make statistics to evaluate nominal values
of those features.
2. We compare features of each attack segment to the previously identified nominal values.
3. We use the tuple notions and duplicated to analyze, as we know the value range for each descriptor.</p>
          <p>We considered an attack as original only if its features are unique.</p>
          <p>To check our results, we analyze segment, without knowing their classification by our system, and
then, we compare if the manual analysis and the output of the system matches (original only detected
by our model, original detected by both model, and original only detected by state-of-the-art model).
1. For 45 original attack only detected by our model, we actually confirm after analysis 32 originals
attacks,
2. For 55 original attack detected by our model and the state-of-the-art models, we confirm after
analysis 44 originals attacks,
3. For 100 original attack detected by state-of-the-art model, we confirm after analysis 88 originals
attack</p>
          <p>To illustrate our analysis, we choose segment number 81 557, a neptune attack, as example: our
attention was drawn by three variations on the descriptors:
1. Duration: The length of time duration of the connection is about 2 seconds, when a duration for
a normal neptune attacks is 0 seconds.
2. Flag: The status of the connection is SF (Normal establishment and termination), when normal
neptune attacks have a flag S0 (connection attempt seen but no reply).
3. Src_bytes: The number of data bytes transferred from source to destination in single connection
is 10714 bytes when a normal neptune attacks is 0.</p>
          <p>For probes attacks, we can mainly cite variation on the service (Destination network service used)
descriptor, like supdup, which appears on the segment number 41 594 (attack nmap), when normal
nmap attacks never use it.</p>
          <p>We also examine segments detected by the state of the art methods as original, in contrast to our model.
The traditional method classifies it as original, when it is just a false positive. For segments number 253
and 125876 which do not present any variation of the patterns, it is inexplicable why they have been
classified so. Finally, segment number 125694 is classified as original, when is rejected (Flag REJ) : it
value are diferent from the normal ones, because the attack never succeed.</p>
          <p>This analysis, carried out on a representative number of samples, confirms that our system is
performing well, as 76% of attacks detected were efectively original : our system can complement
existing methods, contributing in the long term to reducing the number of alerts raised by SOCs
requiring human analysis.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusion and future work</title>
      <p>In this paper, we proposed a method for attack originality detection. The proposed method is based on
the collaboration of diferent weak classifiers with a voting system which is responsible for detecting
non-consensual attacks identified as original. The obtained results are promising, making evidence of
situations where classical techniques fail to detect attack originality. Conclusions have also confirmed
by a CS Group cyber analyst. Results confirm that although competing techniques may detect more
abnormal attacks, they fail to detect others that have been confirmed as original by the expert, confirming
the specificity of our method, and the complement it ofer to state-of-the-art method.
This method is a first step to improve SIEM capabilities of early stage APT detection. Future work will
focus on adding state-of-the-art classifiers to strengthen our voting system, and then investigating deep
learning methods for attack originality detection.</p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgments</title>
      <p>The authors would like to thank KOUAMEN Josee Alvine, analyst of the BU Cyber of CS group, for
the results validation, and the Data &amp; Process Intelligence Team of CS-Group for their precious help,
particularly MENIODEM DELIOTA Keline.
The author(s) have not employed any Generative AI tools.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>D. S.</given-names>
            <surname>Berman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. L.</given-names>
            <surname>Buczak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. S.</given-names>
            <surname>Chavis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. L.</given-names>
            <surname>Corbett</surname>
          </string-name>
          ,
          <article-title>A Survey of Deep Learning Methods for Cyber Security</article-title>
          , https://www.mdpi.com/2078-2489/10/4/122,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>W. A.</given-names>
            <surname>Yousef</surname>
          </string-name>
          , I. Traoré, W. Briguglio, Classifier Calibration:
          <article-title>With Application to Threat Scores in Cybersecurity</article-title>
          ,
          <source>IEEE Transactions on Dependable and Secure Computing</source>
          <volume>20</volume>
          (
          <year>2023</year>
          )
          <fpage>1994</fpage>
          -
          <lpage>2010</lpage>
          . doi:
          <volume>10</volume>
          .1109/TDSC.
          <year>2022</year>
          .
          <volume>3170011</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>O. L.</given-names>
            <surname>Soh</surname>
          </string-name>
          ,
          <article-title>Advanced Persistent Threat Detection Using Anomaly Score Calibration and Multi-class Classification, Master's thesis</article-title>
          , University of Victoria,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Haribalaji</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Ranjana</surname>
          </string-name>
          ,
          <article-title>Distributed Denial of Service (DDOS) Attack Detection Using Classification Algorithm</article-title>
          , in: 2024
          <source>International Conference on Advances in Data Engineering and Intelligent Computing Systems (ADICS)</source>
          ,
          <year>2024</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          . doi:
          <volume>10</volume>
          .1109/ADICS58448.
          <year>2024</year>
          .
          <volume>10533510</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Meira</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Andrade</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Praça</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Carneiro</surname>
          </string-name>
          , G. Marreiros,
          <article-title>Comparative Results with Unsupervised Techniques in Cyber Attack Novelty Detection</article-title>
          , in: P. Novais,
          <string-name>
            <given-names>J. J.</given-names>
            <surname>Jung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. Villarrubia</given-names>
            <surname>González</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Fernández-Caballero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Navarro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>González</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Carneiro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pinto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. T.</given-names>
            <surname>Campbell</surname>
          </string-name>
          , D. Durães (Eds.),
          <source>Ambient Intelligence - Software</source>
          and Applications -,
          <source>9th International Symposium on Ambient Intelligence</source>
          , Springer International Publishing, Cham,
          <year>2019</year>
          , pp.
          <fpage>103</fpage>
          -
          <lpage>112</lpage>
          . doi:
          <volume>10</volume>
          .1007/ 978-3-
          <fpage>030</fpage>
          -01746-0_
          <fpage>12</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>M.</given-names>
            <surname>Tavallaee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Bagheri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Ghorbani</surname>
          </string-name>
          ,
          <article-title>A detailed analysis of the KDD CUP 99 data set</article-title>
          ,
          <source>in: 2009 IEEE Symposium on Computational Intelligence for Security and Defense Applications</source>
          ,
          <year>2009</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          . doi:
          <volume>10</volume>
          .1109/CISDA.
          <year>2009</year>
          .
          <volume>5356528</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>I. Trrad</surname>
          </string-name>
          ,
          <article-title>Applying Deep Learning Techniques for Network Trafic Classification: A Comparison Study on the NSL-KDD</article-title>
          <string-name>
            <surname>Dataset</surname>
          </string-name>
          ,
          <year>2024</year>
          . doi:
          <volume>10</volume>
          .21203/rs.3.rs-
          <volume>3869444</volume>
          /v1.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A. D.</given-names>
            <surname>Vibhute</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. H.</given-names>
            <surname>Patil</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. V.</given-names>
            <surname>Mane</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. V.</given-names>
            <surname>Kale</surname>
          </string-name>
          ,
          <article-title>Towards Detection of Network Anomalies using Machine Learning Algorithms on the NSL-KDD Benchmark Datasets</article-title>
          ,
          <source>Procedia Computer Science</source>
          <volume>233</volume>
          (
          <year>2024</year>
          )
          <fpage>960</fpage>
          -
          <lpage>969</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.procs.
          <year>2024</year>
          .
          <volume>03</volume>
          .285.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>D.</given-names>
            <surname>Protic</surname>
          </string-name>
          , Review of KDD Cup '
          <volume>99</volume>
          ,
          <string-name>
            <surname>NSL-KDD</surname>
          </string-name>
          and
          <article-title>Kyoto 2006+ datasets</article-title>
          ,
          <source>Vojnotehnicki glasnik 66</source>
          (
          <year>2018</year>
          )
          <fpage>580</fpage>
          -
          <lpage>596</lpage>
          . doi:
          <volume>10</volume>
          .5937/vojtehg66-
          <fpage>16670</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>R.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Grover</surname>
          </string-name>
          ,
          <article-title>Enhancing Cybersecurity with Machine Learning: Evaluating the Eficacy of Isolation Forests and Autoencoders in Anomaly Detection</article-title>
          ,
          <source>in: 2024 7th International Conference on Circuit Power and Computing Technologies (ICCPCT)</source>
          , volume
          <volume>1</volume>
          ,
          <year>2024</year>
          , pp.
          <fpage>1017</fpage>
          -
          <lpage>1021</lpage>
          . doi:
          <volume>10</volume>
          . 1109/ICCPCT61902.
          <year>2024</year>
          .
          <volume>10673338</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>R. C.</given-names>
            <surname>Ripan</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Islam</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Alqahtani</surname>
            ,
            <given-names>I. H.</given-names>
          </string-name>
          <string-name>
            <surname>Sarker</surname>
          </string-name>
          ,
          <article-title>Efectively predicting cyber-attacks through isolation forest learning-based outlier detection</article-title>
          ,
          <source>SECURITY AND PRIVACY 5</source>
          (
          <year>2022</year>
          )
          <article-title>e212</article-title>
          . doi:
          <volume>10</volume>
          . 1002/spy2.
          <fpage>212</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>N. J.</given-names>
            <surname>Johannesen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. L.</given-names>
            <surname>Kolhe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Goodwin</surname>
          </string-name>
          ,
          <article-title>Vertical Approach Anomaly Detection Using Local Outlier Factor</article-title>
          , in: H.
          <string-name>
            <surname>Haes Alhelou</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Hatziargyriou</surname>
            ,
            <given-names>Z. Y.</given-names>
          </string-name>
          <string-name>
            <surname>Dong</surname>
          </string-name>
          (Eds.),
          <source>Power Systems Cybersecurity: Methods, Concepts, and Best Practices</source>
          , Springer International Publishing, Cham,
          <year>2023</year>
          , pp.
          <fpage>297</fpage>
          -
          <lpage>310</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>031</fpage>
          -20360-2_
          <fpage>12</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>A. R.</given-names>
            <surname>Vasudevan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Selvakumar</surname>
          </string-name>
          ,
          <article-title>Local outlier factor and stronger one class classifier based hierarchical model for detection of attacks in network intrusion detection dataset</article-title>
          ,
          <source>Frontiers of Computer Science</source>
          <volume>10</volume>
          (
          <year>2016</year>
          )
          <fpage>755</fpage>
          -
          <lpage>766</lpage>
          . doi:
          <volume>10</volume>
          .1007/s11704-015-5116-8.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>