<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>An Isolation Forest - based approach for brute force attack detection</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Olha Mykhaylova</string-name>
          <email>olha.o.mykhailova@lpnu.ua</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andriy Shtypka</string-name>
          <email>andriyko7788@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Taras Fedynyshyn</string-name>
          <email>fedynyshyn.taras@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Lviv Polytechnic National University</institution>
          ,
          <addr-line>12 Stepan Bandera str., Lviv, 79000</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In today's rapidly evolving digital landscape, characterized by dynamic cyber threats, brute force attacks persist as among the most prevalent and enduring security challenges. In the face of evolving cyber-attacks, conventional detection techniques frequently demonstrate inadequacies, prompting the exploration of innovative solutions like the integration of machine learning algorithms. This paper introduces a novel intelligent model utilizing decision trees, designed to detect anomalies in user behavior indicative of potential brute force attacks. Additionally, this paper includes the Python source code for implementing the model as well as presenting the obtained results. The paper explores the effectiveness of an intelligent decision tree-based model in detecting brute force attacks during system logins. It delves into the algorithmic underpinnings of these models, highlights their advantages over traditional detection methods, and examines practical considerations for their implementation and utilization. The proposed Isolation Forest-based model effectively detects brute force attacks with high accuracy, adaptability, and reduced false positives, but it requires careful tuning of anomaly thresholds, faces limitations in highly imbalanced datasets where attack instances are rare, and may be vulnerable to sophisticated adversarial tactics that mimic normal behavior, highlighting the need for further improvements in its robustness and sensitivity.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;intrusion detection model</kwd>
        <kwd>Isolation Forest</kwd>
        <kwd>brute force</kwd>
        <kwd>machine learning1</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>In the modern digital world, where cyber threats evolve at an incredible pace, brute force attacks
continue to remain one of the most common and persistent threats. These attacks, which
involve the continuous guessing of user credentials, pose a serious threat to the security of
systems and data. However, traditional methods of detection often prove to be insufficiently
effective in the changing landscape of cyber-attacks, opening the door to innovative approaches
such as the application of machine learning algorithms.</p>
      <p>One such approach is the development of intelligent models based on decision trees, capable
of identifying anomalies in user behavior that may indicate attempts of brute force attacks.
These models, able to analyze large volumes of data and promptly detect potential threats,
represent a significant advancement in countering these attacks. They not only enhance the
accuracy of attack detection but also reduce the number of false positives, optimizing the
performance of security teams.</p>
      <p>The intelligent model for detecting brute force attacks is based on decision-making using
decision trees as a machine learning algorithm to classify network traffic as either normal or
anomalous. The model applies the decision tree algorithm for training based on a labeled dataset
of network traffic, which includes examples of both successful login attempts or unsuccessful
attempts within acceptable norms, as well as anomalous unsuccessful attempts that may
indicate malicious unauthorized access attempts. During training, the model identifies criteria
by which brute force attacks are most likely to occur.</p>
      <p>Upon completion of training, the model can be utilized for classifying new network traffic,
such as login attempts into the system. If the model identifies a segment of traffic as malicious, it
is likely that this traffic is part of a brute force attack. This information can serve as the basis for
taking actions such as blocking the traffic or notifying the network administrator.</p>
      <p>Decision trees represent a type of machine learning algorithm that is optimal for
classification tasks. One of the advantages of using decision trees for detecting brute force
attacks is their high effectiveness in recognizing attacks that are not yet known to the system,
thanks to their ability to learn from new data and adapt to changes in the attack landscape. It is
also important to note that decision trees can be easily interpreted, facilitating understanding of
the decisions made by the model.</p>
      <p>However, there are some considerations when using decision trees for detecting brute force
attacks. One such consideration is that decision trees require continual training to function
correctly. Additionally, decision trees need to be adapted individually for each project,
considering its specific characteristics and security requirements.</p>
      <p>To enhance the accuracy and effectiveness of the intelligent intrusion detection model for
system login, it is crucial to conduct performance evaluation of brute force attack detection
using decision trees. Effective evaluation of brute force attack detection models is vital for
ensuring system security. Decision trees, with their capability to identify complex patterns in
data, offer a promising approach for this task.</p>
      <p>The following metrics are used to evaluate the effectiveness of decision tree models in the
context of brute force attack detection:
 True Positive Rate (TPR), also known as recall, measures the proportion of actual brute
force attacks that were correctly identified. A high TPR indicates that the model
effectively detects true attacks.
 False Positive Rate (FPR) assesses the proportion of normal login attempts that are
incorrectly classified as brute force attacks. A low FPR ensures that the model minimizes
unnecessary alarms.
 Accuracy measures the proportion of login attempts labeled as brute force attacks that
are actually true attacks. A high accuracy value indicates that the model is precise in its
classifications.</p>
      <p>In addition to these metrics, it is also important to evaluate the following characteristics of
decision tree models:
 Detection time is a critical factor in real-time attack detection scenarios. Swift detection
ensures that the system can promptly respond to an attack.
 Computational efficiency is also an important characteristic of decision tree models,
especially when processing large volumes of data.
 Resistance to adversarial attacks is an important characteristic for any brute force attack
detection model. Adversaries may employ various methods to evade detection, such as
using unique login credentials or introducing subtle variations into their attack patterns.
 Interpretability of decision tree models is important for ensuring their credibility and
interpretability. Interpretability methods can provide insight into the decision-making
process of the model, allowing security analysts to verify its performance and identify
potential biases.</p>
      <p>Each performance evaluation metric corresponds to a specific task. For example, TPR is
important to ensure that the model does not miss real attacks. FPR is important to ensure that
the model does not generate too many false positives.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Works</title>
      <p>
        In brute-force attacks, the assailant systematically submits every conceivable value as inputs for
account credentials to gain unauthorized access to the system's account data. These attacks are
characterized by two primary methodologies: dictionary attacks, which exhaustively test all
entries in a predefined list, and random sequence methods, which systematically test all feasible
string combinations in a sequential manner [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. A prevalent strategy for mitigating brute-force
attacks involves implementing access controls triggered by repeated incorrect password entries.
For servers, this entails establishing a threshold for login failures via an account lock policy,
which restricts access to the account once the threshold is surpassed [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>This research paper delves into the creation of an Intrusion Detection System (IDS),
specifically brute force attacks, with particular emphasis on the IDS's ability to withstand
adversarial attacks and the reliability of explainable AI.</p>
      <p>
        Over the past decade, numerous intrusion detection Systems have been developed to
safeguard cyber networks against malicious attacks [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Notably, Machine Learning (ML)-based
IDS have demonstrated outstanding performance owing to their capacity to learn vast numbers
of parameters. However, these sophisticated models, often termed black-box models, lack
interpretability [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], which contradicts the essential need for transparent decision-making in IDS
operations. Given that even a single erroneous prediction by an IDS can expose the system to
significant cyber threats, the integration of eXplainable Artificial Intelligence (XAI) is
imperative in traditional IDS frameworks to enhance credibility and reliability.
      </p>
      <p>
        Mane et al. [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] employed the NSL-KDD dataset alongside a Deep Neural Network
(DNN)based Machine Learning (ML) model for network intrusion detection. In an effort to enhance
transparency, they employed five distinct XAI frameworks to illustrate the behavior of the
trained model. Nevertheless, they did not leverage the explanations provided by any XAI
framework to validate the credibility of the predicted outcomes.
      </p>
      <p>
        Mahbooba et al. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], similarly, addressed the explanation of individual predicted outcomes by
deriving rules from the decision tree trained and assessed on the KDD dataset. These rules were
exclusively employed to clarify each predicted outcome and the overall model response.
However, their focus did not extend to adversarial attacks or enhancing Intrusion Detection
Systems (IDS) using explanations from XAI tools
      </p>
      <p>
        Fidel et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] introduced a framework based on XAI signatures to distinguish between
adversarial samples and normal network traffic. They assessed their approach using datasets
typically employed in image recognition tasks and achieved an accuracy of approximately 97%
in detecting adversarial attacks. Such defense mechanisms are essential in cyber networks.
Hence, this paper suggests a novel intrusion detection approach aimed at verifying the
credibility of machine learning model predictions and ensuring superior performance in both
normal and adversarial scenarios. Moreover, it enhances transparency in the decision-making
process, thereby bolstering user trust.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Proposed model</title>
      <p>Isolation Forest is an innovative unsupervised learning method specifically designed for
anomaly detection in data. A key feature of this method is its ability to effectively detect
anomalies without the need for prior definition or labeling of normal data. Isolation Forest
utilizes an ensemble of randomized decision trees to "isolate" observations, whereby anomalous
data points are typically isolated in fewer steps than normal observations.</p>
      <p>The key idea of the algorithm lies in the fact that anomalies typically represent a minority in
the dataset and exhibit differences from normal observations, allowing them to be isolated more
quickly. The process of partitioning the data is performed recursively until each instance is
isolated in its own "leaf" of the tree. The number of splits required to isolate an instance is used
as a measure of its anomaly score.</p>
      <p>Isolation Forest demonstrates high effectiveness in various applications, including:
 Fraud detection: Identifying anomalous transactions or user behavior that may indicate
financial fraud.
 Intrusion detection systems: Detecting anomalous patterns in network traffic that may
indicate unauthorized access attempts.
 Recommendation systems: Identifying non-standard user behavior or anomalous
purchasing patterns that may impact recommendation algorithms.
 Detection of anomalies in large datasets: Identifying anomalous records in large datasets
that may indicate data errors, external interference, or other unforeseen phenomena.</p>
      <p>The underlying principle of the Isolation Forest algorithm is that anomalous data points are
more readily distinguishable from the rest of the dataset. To isolate a data point, the algorithm
iteratively creates partitions within the dataset by randomly choosing an attribute and
subsequently selecting a split value at random between the minimum and maximum values
permitted for that attribute.</p>
      <p>The process of recursive partitioning can be visualized through a tree structure called the
Isolation Tree. The number of partitions needed to isolate a point can be understood as the
length of the path within this tree, starting from the root and terminating at a leaf node. For
instance, in Figure 1, the path length of point  is longer than that of  in Figure 2.</p>
      <p>
        Let  = {1, …, } be [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] a set of d-dimensional points and ′⊂ . An Isolation Tree (iTree) is
defined as a data structure with the following properties:
1. for each node  in the Tree,  is either an external-node with no child, or an
internalnode with one “test” and exactly two child nodes ( and ).
2. a test at node  consists of an attribute  and a split value  such that the test  &lt; 
determines the traversal of a data point to either  or .
      </p>
      <p>To build an iTree, the algorithm recursively divides ′ by randomly selecting an attribute 
and a split value , until either:
1. the node has only one instance, or
2. all data at the node have the same values.</p>
      <p>When the iTree is fully grown, each point in  is isolated at one of the external nodes.
Intuitively, the anomalous points are those (easier to isolate, hence) with the smaller path length
in the tree, where the path length ℎ() of point  ∈  is defined as the number of edges 
traverses from the root node to get to an external node.</p>
      <p>To detect anomalies using IForest, two main steps need to be executed:
1. Training the model. The model is built by taking a partial sample of the dataset.
2. Testing the model. The test dataset is fed into the model to compute anomaly scores for
each point.</p>
      <p>The diagram in Figure 3 illustrates the algorithmic steps of applying the IForest model.</p>
      <p>The IForest algorithm can be used to detect brute force attacks on a login system. If an
attacker attempts to guess the password, they will make many incorrect login attempts. The
IForest algorithm can detect and label such attempts as anomalies. The process of applying the
Isolation Forest algorithm for a login system attack detection model consists of several steps.</p>
      <p>The first step is to gather data about login attempts to the system. This data may include
characteristics such as:
 IP address
 Username
 Password
 Date</p>
      <p>The second step is to create an IForest model. This can be done using the Isolation Forest
library from the scikit-learn package.</p>
      <p>The third step is to train the IForest model on the gathered data. This process may take some
time depending on the size of the dataset.</p>
      <p>After training the IForest model, it's necessary to determine the anomaly threshold. This
threshold value determines which login attempts are considered anomalous.</p>
      <p>The fifth step is to use the IForest model to assess anomalies for new login attempts. If a login
attempt has an anomaly score lower than the anomaly threshold, it is considered anomalous.</p>
      <p>Brute force attacks on login systems typically have a short path to the root of the tree. This is
because the attacker attempting to crack the password will try many incorrect passwords. If a
login attempt has an anomaly score below the anomaly threshold, it is considered anomalous.
This login attempt can be classified as a brute force attack. The anomaly threshold can be
optimized to achieve an optimal balance between sensitivity and specificity. Sensitivity is the
probability that the model correctly classifies an anomalous login attempt as an anomaly.
Specificity is the probability that the model correctly classifies normal login attempts as normal.
A too low anomaly threshold will result in high sensitivity but also a high number of false
positive results. A too high anomaly threshold will result in high specificity but also a high
number of false negative results.</p>
      <p>IForest is an effective method for detecting brute force attacks on login systems. It has several
advantages, such as:
 Speed
 Accuracy
 Robustness to noise
 Adaptability to changes</p>
    </sec>
    <sec id="sec-4">
      <title>4. Results and discussion</title>
      <p>In this section, the Python code implementation is demonstrated to create an intelligent
intrusion detection system (for brute force attacks) on the login system. After displaying each
written part of the code, the execution result and its description is provided.</p>
      <p>The working principle of the intelligent system model involves detecting unsuccessful login
attempts to specific accounts within a system or software application based on login history and
providing information about each anomaly (users and anomalous IP addresses). This enables the
individual or the monitoring model itself to take appropriate measures to ensure integrity,
confidentiality, and availability.</p>
      <p>Before writing the code, we need to first obtain and save the login data into a CSV file for
training and applying the model. The CSV file should have 4 columns (IP-address, user, time,
login result), containing data about the IP addresses from which login attempts were made, the
users whose accounts were attempted to be logged into, the time, and the login result, as shown
in Figure 4.</p>
      <p>
        The next step is to install and utilize the Scikit-learn library along with Isolation Forest for
training and application, Pandas for file operations, NumPy for numerical functions, and
Matplotlib for creating plots. See Appendix A for references [
        <xref ref-type="bibr" rid="ref10 ref11 ref12">10, 11, 12</xref>
        ].
      </p>
      <p>To filter the data, we create two DataFrames, failed logins and success logins, where we store
the data and count the number of failed and successful login attempts, respectively. Then, we
merge them into one DataFrame, grouped data, where we store data about users, IP addresses,
and the count of failed and successful logins.</p>
      <p>To train the anomaly detection model for the login system, we need to utilize the grouped
data DataFrame containing information about the count of failed and successful login attempts.
For this purpose, we create our model (model) and specify parameters (contamination='auto',
which denotes the percentage of anomalous points from the total by isolating normal points
from anomalies, random state=42, defining the initial state of the random number generator).
Then, based on the count of failed and successful attempts for each IP address, the model
predicts whether there is anomaly, labeling each row accordingly: -1 if an anomaly is detected
and 1 if not, as showed on Figure 5.</p>
      <p>To visually represent the model's output, we need to use the Matplotlib library. The plot is
constructed based on information about failed attempts and is displayed corresponding to each
user from the "Logs.csv" file. The visualized results presented on Figure 6.</p>
      <p>To obtain information about suspicious IP addresses from which a specific user experienced
an anomalous number of login attempts, you need to filter IP addresses from the DataFrame
“grouped_data” based on the "anomaly" index and present it as a message.</p>
      <p>The development of an intelligent intrusion detection system for identifying brute force
attacks on login systems demonstrates significant potential in ensuring digital security. The
application of the Isolation Forest algorithm in the context of unsupervised learning enables
effective identification of anomalies in large datasets, particularly in login histories.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Discussion</title>
      <p>The proposed Isolation Forest-based approach effectively detects brute force attacks with
high accuracy and reduced false positives compared to traditional methods. Its ability to
isolate anomalous login attempts in real-time enhances its utility in large-scale systems,
offering both speed and adaptability as it learns from evolving data patterns. One of the
model’s major strengths is its unsupervised nature, allowing it to identify anomalies
without the need for labeled datasets, making it suitable for dynamic environments.
However, the model’s performance heavily depends on setting an optimal anomaly
threshold, which requires fine-tuning to balance sensitivity and specificity. A key limitation
is that in cases of highly imbalanced datasets, where brute force attacks are rare, the model
may struggle with sensitivity, potentially missing some attacks. Additionally, the model can
be vulnerable to adversarial attacks that are designed to mimic normal login behavior,
reducing its effectiveness in more sophisticated threat scenarios. Despite these limitations,
the Isolation Forest method shows promise as a scalable and efficient solution for brute
force detection. Future work could explore hybrid approaches or enhancements to further
improve the model’s resilience against adversarial tactics and improve its performance in
imbalanced data sets.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>The developed algorithm demonstrates high effectiveness in anomaly detection, especially in
contexts where anomalies are relatively rare compared to the overall number of observations.
Its ability to quickly isolate anomalous data points makes it an ideal tool for detecting brute
force attack attempts.</p>
      <p>
        Usage of this method may automate the process of identifying suspicious login attempts,
reducing the need for constant monitoring by security analysts and increasing the speed of
response to potential threats. This approach can be utilized on web and mobile application
services and may be highly valuable if applied on critical infrastructure computer systems [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
The unsupervised learning usage helps reduce the number of false positives, as the model adapts
to changes in data behavior, ensuring more accurate detection of real anomalies.
      </p>
      <p>The developed method seamlessly integrates with existing login systems and databases,
enabling the quick implementation of an additional layer of security without the need for
significant modifications to the existing infrastructure. It also provides information about users
and IP addresses associated with anomalous access attempts, which can be used for further
analysis and the development of countermeasures.</p>
    </sec>
    <sec id="sec-7">
      <title>A. Algorithm implementation python code</title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Park</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gupta</surname>
            ,
            <given-names>B.B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Park</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          (
          <year>2021</year>
          ).
          <article-title>Network log-based SSH brute-force attack detection model</article-title>
          .
          <source>Computers, Materials &amp; Continua</source>
          ,
          <volume>68</volume>
          (
          <issue>1</issue>
          ),
          <fpage>887</fpage>
          -
          <lpage>901</lpage>
          . https://doi.org/10.32604/cmc.
          <year>2021</year>
          .
          <volume>015172</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>P. S.</given-names>
            <surname>Abril</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Plant</surname>
          </string-name>
          ,
          <article-title>The patent holder's dilemma: Buy, sell, or troll?</article-title>
          ,
          <source>Communications of the ACM</source>
          <volume>50</volume>
          (
          <year>2007</year>
          )
          <fpage>36</fpage>
          -
          <lpage>44</lpage>
          . https://doi.org/10.1145/1188913.1188915.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Abid</given-names>
            <surname>Salih</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            and
            <surname>Abdulazeez</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.M.</surname>
          </string-name>
          (
          <year>2021</year>
          )
          <article-title>“Evaluation of Classification Algorithms for Intrusion Detection System: A Review”</article-title>
          ,
          <source>Journal of Soft Computing and Data Mining</source>
          ,
          <volume>2</volume>
          (
          <issue>1</issue>
          ), pp.
          <fpage>31</fpage>
          -
          <lpage>40</lpage>
          . Available at: https://publisher.uthm.edu.my/ojs/index.php/jscdm/article/view/7982 (Accessed: 29
          <source>April</source>
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Rai</surname>
            ,
            <given-names>A. Explainable</given-names>
          </string-name>
          <article-title>AI: from black box to glass box</article-title>
          .
          <source>J. of the Acad. Mark. Sci</source>
          .
          <volume>48</volume>
          ,
          <fpage>137</fpage>
          -
          <lpage>141</lpage>
          (
          <year>2020</year>
          ). https://doi.org/10.1007/s11747-019-00710-5.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Mane</surname>
          </string-name>
          and
          <string-name>
            <given-names>D.</given-names>
            <surname>Rao</surname>
          </string-name>
          , “
          <article-title>Explaining network intrusion detection system using explainable ai framework</article-title>
          ,
          <source>” arXiv preprint arXiv: 2103.07110</source>
          ,
          <year>2021</year>
          . https://doi.org/10.48550/arXiv.2103.07110.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Mahbooba</surname>
            ,
            <given-names>Basim</given-names>
          </string-name>
          &amp; Timilsina, Mohan &amp; Sahal, Radhya &amp; Serrano,
          <string-name>
            <surname>Martin.</surname>
          </string-name>
          (
          <year>2021</year>
          ).
          <article-title>Explainable Artificial Intelligence (XAI) to Enhance Trust Management in Intrusion Detection Systems Using Decision Tree Model</article-title>
          .
          <source>Complexity</source>
          .
          <year>2021</year>
          .
          <volume>11</volume>
          . https://doi.org/10.1155/
          <year>2021</year>
          /6634811.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>G.</given-names>
            <surname>Fidel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Bitton</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Shabtai</surname>
          </string-name>
          , “
          <article-title>When explainability meets adversarial learning: detecting adversarial examples using shap signatures,”</article-title>
          <source>in Proceedings of 2020 International Joint Conference on Neural Networks (IJCNN)</source>
          , Glasgow, UK,
          <year>2020</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Isolation</surname>
            <given-names>forest</given-names>
          </string-name>
          , Available at: https://en.wikipedia.org/wiki/Isolation_forest.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Zou</surname>
            ,
            <given-names>Zhuping</given-names>
          </string-name>
          &amp; Xie, Yulai &amp; Huang, Kai &amp; Xu,
          <string-name>
            <surname>Gongming</surname>
          </string-name>
          &amp; Feng,
          <string-name>
            <surname>Dan</surname>
          </string-name>
          &amp; Long,
          <string-name>
            <surname>Darrell.</surname>
          </string-name>
          (
          <year>2019</year>
          ).
          <article-title>A Docker Container Anomaly Monitoring System Based on Optimized Isolation Forest</article-title>
          .
          <source>IEEE Transactions on Cloud Computing. PP. 1-1.</source>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Python</surname>
            <given-names>documentation</given-names>
          </string-name>
          , Available at: https://www.python.org/doc/ (Accessed: 11
          <source>April</source>
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <article-title>Scikit-learn Isolation Forest documentation</article-title>
          . Available at: https://scikit-learn.org/stable/modules/generated/sklearn.ensemble.IsolationForest.html
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <source>[12] Matplotlib 3.8</source>
          .4 documentation. Available at: https://matplotlib.org/stable/index.
          <source>html (Accessed: 11 April</source>
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Olha</surname>
            <given-names>Mykhaylova</given-names>
          </string-name>
          , Taras Fedynyshyn, Anastasiia Datsiuk, Bohdan Fihol, Hennadii Hulak:
          <article-title>Mobile Application as a Critical Infrastructure Cyberattack Surface</article-title>
          .
          <source>Proceedings of the Cybersecurity Providing in Information and Telecommunication Systems II co-located with International Conference on Problems of Infocommunications. Science and Technology (PICST</source>
          <year>2023</year>
          ), Kyiv, Ukraine, October
          <volume>26</volume>
          ,
          <year>2023</year>
          , CEUR-WS.org/Vol-
          <volume>3550</volume>
          , urn:nbn:de:
          <fpage>0074</fpage>
          -
          <lpage>3550</lpage>
          -0.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>