=Paper= {{Paper |id=Vol-3922/paper7 |storemode=property |title=Advancing Cybersecurity with LLMs: A Comprehensive Review of Intrusion Detection Systems and Emerging Application |pdfUrl=https://ceur-ws.org/Vol-3922/paper7.pdf |volume=Vol-3922 |authors=Djallel Hamouda,Mohamed Amine Ferrag,Nadjette Benhamida,Hamid Seridi |dblpUrl=https://dblp.org/rec/conf/iam/HamoudaFBS24 }} ==Advancing Cybersecurity with LLMs: A Comprehensive Review of Intrusion Detection Systems and Emerging Application== https://ceur-ws.org/Vol-3922/paper7.pdf
                         Advancing Cybersecurity with LLMs: A Comprehensive
                         Review of Intrusion Detection Systems and Emerging
                         Applications
                         Hamouda Djallel1 , Mohamed Amine Ferrag1 , Benhamida Nadjette1 and Seridi Hamid1
                         1
                             Labstic Laboratory, Department of Computer Science, Guelma University, B.P. 401, 24000, Guelma, Algeria


                                        Abstract
                                        The rapid advancements in Transformers and Large Language Models (LLMs) have significantly transformed
                                        the landscape of cybersecurity, particularly in Intrusion Detection Systems (IDS). These models offer enhanced
                                        detection accuracy, scalability, and adaptability, surpassing traditional approaches in identifying and mitigating
                                        sophisticated cyber threats. This survey explores the integration of LLMs in IDS by addressing six key research
                                        dimensions: foundational methodologies, comparative performance with classical techniques, challenges in
                                        interpretability, practical applications, emerging trends, and directions for future research. By synthesizing the
                                        latest advancements, this work aims to provide a comprehensive framework for understanding the role of LLMs
                                        in strengthening cybersecurity and fostering innovation in network security and anomaly detection.

                                        Keywords
                                        Large Language Models, Transformers, Intrusion Detection Systems, Cybersecurity, Malware Detection




                         1. Introduction
                         The rapid digital transformation of industries and societies has significantly increased reliance on
                         interconnected systems, making cybersecurity a critical concern [1]. With the increasing volume, diver-
                         sity, and sophistication of cyber threats, traditional Intrusion Detection Systems (IDS), including those
                         employing machine learning (ML) and deep learning (DL) techniques, face limitations in adaptability
                         and resilience [2]. Although ML and DL-based IDS have enhanced detection accuracy and adaptability
                         compared to static, signature-based systems, they often struggle with challenges such as handling
                         high-dimensional data, evolving attack vectors, and adversarial inputs. These constraints underscore
                         the need for more advanced and robust approaches to counter modern cyber threats [3].
                            Recent advancements in Artificial Intelligence (AI), particularly in Transformers [4] and Large
                         Language Models (LLMs), have revolutionized numerous domains. Initially developed for natural
                         language processing tasks, LLMs like BERT [5] and GPT [6] excel at identifying complex patterns
                         and contextual relationships within extensive datasets. These attributes make LLMs a promising
                         enhancement to existing IDS by addressing limitations in scalability, adaptability, and the detection of
                         novel threats that traditional ML/DL-based approaches struggle to manage effectively [7].
                            This survey aims to explore the integration of LLMs into IDS, providing a comprehensive review of
                         methodologies, applications, challenges, and future directions. Specifically, the paper addresses the
                         following objectives:
                                 • Examine the foundational methodologies enabling the deployment of LLMs in IDS.
                                 • Compare the performance of LLM-based systems with traditional IDS approaches, highlighting
                                   their strengths and limitations.
                         Proceedings of the International IAM’24: International Conference on Informatics and Applied Mathematics, December 04–05,
                         2024, Guelma, Algeria
                         ∗
                             Corresponding author.
                         †
                             These authors contributed equally.
                         Envelope-Open hamouda.djallel@univ-guelma.dz (H. Djallel); ferrag.mohamedamine@univ-guelma.dz (M. A. Ferrag);
                         benhamida.nadjette@univ-guelma.dz (B. Nadjette); seridi.hamid@univ-guelma.dz (S. Hamid)
                         Orcid 0000-0003-2168-4192 (H. Djallel); 0000-0002-0632-3172 (M. A. Ferrag); 0000-0002-5540-8594 (B. Nadjette);
                         0000-0002-0236-8541 (S. Hamid)
                                        © 2024 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).


CEUR
                  ceur-ws.org
Workshop      ISSN 1613-0073
Proceedings
    • Discuss key challenges, including computational complexity, dataset limitations, and interpretabil-
      ity concerns.
    • Highlight practical applications of LLMs in cybersecurity, such as network security, malware
      detection, and phishing prevention.
    • Propose future research directions to overcome current limitations and drive further advancements
      in the field.

By consolidating recent advancements and delivering a structured analysis, this study aims to equip
researchers and practitioners with a roadmap for deploying LLMs to develop more efficient, adaptable,
and comprehensive cybersecurity solutions


2. Background and Preliminaries
This section outlines foundational concepts, including LLMs, IDS, and their intersection, to set the stage
for subsequent discussions

2.1. Large Language Models and Transformers
Transformers have reshaped artificial intelligence, with architectures such as BERT , GPT , and T5
achieving remarkable performance across diverse tasks [8]. Their foundational self-attention mechanism
enables efficient modeling of relationships between input elements, allowing them to capture long-range
dependencies effectively.
   LLMs, built on these architectures, leverage extensive pre-training on large corpora followed by
fine-tuning for specific tasks. Key advantages of LLMs include:

    • Pattern Recognition: The ability to identify complex patterns within large and heterogeneous
      datasets [6].
    • Contextual Understanding: Robust handling of sequential and contextual information [5].
    • Scalability: High adaptability to diverse domains, including cybersecurity [8].

These properties make LLMs particularly suited for addressing the challenges posed by modern cyber
threats.

2.2. Intrusion Detection Systems (IDS)
IDS are critical components of cybersecurity infrastructures, designed to monitor, detect, and respond
to suspicious activities within networks or systems [2]. Broadly, IDS are categorized into:

    • Signature-Based IDS: These systems rely on predefined rules and patterns to identify known
      threats. While efficient, they struggle with zero-day attacks and novel threats.
    • Anomaly-Based IDS: These systems use statistical or machine learning techniques to detect
      deviations from normal behavior, making them more adaptable to new threats but prone to higher
      false positive rates.
    • Hybrid IDS: Combining signature-based and anomaly-based approaches, hybrid IDS aim to
      balance accuracy and adaptability.

   Despite advancements, traditional IDS often face challenges, including limited adaptability, high
false positive rates, and computational inefficiency, particularly in handling large-scale or dynamic
environments [2].
2.3. Intersection of LLMs and IDS
Integrating LLMs into IDS effectively addresses key limitations of traditional systems. Through their
advanced pattern recognition and contextual analysis capabilities, LLMs enable IDS to:

    • Enhance Detection Accuracy: Identify complex and evolving threats with greater precision
      [9].
    • Improve Adaptability: Generalize effectively across diverse attack scenarios without extensive
      reconfiguration [7].
    • Reduce False Positives: Provide more reliable threat detection, minimizing unnecessary alerts
      [9].

This synergy between LLMs and IDS represents a paradigm shift in cybersecurity, offering solutions
that are both scalable and robust against sophisticated threats [10].


3. Current Methodologies
The integration of LLMs into IDS systems has led to the development of various methodologies aimed at
enhancing detection capabilities. This section explores the primary approaches, including pre-trained
model customization, fine-tuning strategies, integration into IDS pipelines, and key algorithms and
models [11].

                                              Raw Data
                                            (Logs, Traffic)



                                            Preprocessing
                                  (Tokenization, Feature Extraction)



                                            LLM Analysis
                                          (Threat Detection)



                                           Output (Alerts,
                                            Mitigation)

Figure 1: Workflow of an LLM-Based Intrusion Detection System



3.1. Pre-trained Models and Customization
Pre-trained LLMs, such as BERT and GPT, are trained on extensive corpora, capturing diverse linguistic
patterns. Customization techniques are applied to adapt these models for cybersecurity applications:

    • Domain-Specific Pre-training: Further training LLMs on cybersecurity-related datasets to
      imbue them with domain-specific knowledge [12].
    • Embedding Alignment: Adjusting embeddings to align with cybersecurity terminologies and
      concepts, enhancing the model’s contextual understanding [9].
3.2. Fine-tuning Strategies
Fine-tuning involves adapting pre-trained LLMs to specific tasks within IDS [13] :

    • Supervised Fine-Tuning: Utilizing labeled datasets to train LLMs for tasks like anomaly detec-
      tion and threat classification.
    • Transfer Learning: Applying knowledge from related domains to improve performance on
      cybersecurity tasks, especially when labeled data is scarce [14].

3.3. Integration into IDS Pipelines
Incorporating LLMs into IDS involves designing workflows that leverage their capabilities [11]:

    • Data Preprocessing: Converting network logs and alerts into formats suitable for LLM process-
      ing .
    • Real-Time Analysis: Implementing LLMs to analyze data streams in real-time, enabling prompt
      detection and response.

3.4. Key Algorithms and Models
Several models and algorithms have been developed to enhance IDS using LLMs :

    • BertIDS: A BERT-based model fine-tuned for identifying and classifying network attacks, demon-
      strating improved accuracy over traditional methods [15] [9].
    • HuntGPT: An LLM-powered intrusion detection dashboard that integrates proactive threat
      hunting with explainable AI frameworks [16].
    • ChatIDS: An approach leveraging LLMs to make IDS alerts understandable to non-experts,
      enhancing interpretability and response [17].

These methodologies represent the forefront of integrating LLMs into IDS, offering enhanced detection
capabilities and adaptability to evolving cyber threats.


4. Comparative Analysis with Traditional Techniques
Integrating large language models (LLMs) into intrusion detection systems (IDS) marks a transformative
shift from traditional methodologies. This section presents a comprehensive comparative analysis,
highlighting the strengths and challenges of LLM-based IDS in relation to traditional approaches. The
key differences are summarized in Table 1.
   LLM-based IDS outperform traditional methods in critical metrics such as accuracy and F1-score,
demonstrating superior capability in handling diverse and complex datasets [10] [7] [13] [11]. These
systems offer several advantages over traditional IDSs:

    • Enhanced Scalability: LLMs effectively process large-scale and dynamic datasets, overcoming
      the scalability limitations of traditional systems.
    • Improved Adaptability: With robust generalization capabilities, LLMs are better suited to
      detect novel threats and zero-day attacks.
    • Contextual Analysis: Leveraging natural language understanding, LLM-based IDS derive
      context from logs and alerts, reducing ambiguity in threat identification.
    • Reduced False Positives: Advanced contextual comprehension allows for more accurate threat
      classification, leading to fewer unnecessary alerts.

  These features are contrasted with traditional IDS approaches in Table 1:
  Despite these advantages, LLM-based IDS face notable challenges:
Table 1
Comparison of Traditional IDS and LLM-Based IDS

         Feature                       Traditional IDS                  LLM-Based IDS
         Detection Method              Rule-based    or     anomaly-    Contextual analysis using
                                       based                            deep learning
         Adaptability to New Threats   Limited                          High
         Scalability                   Challenging        for   large   Scalable with proper opti-
                                       datasets                         mization
         False Positives               High                             Lower due to contextual un-
                                                                        derstanding
         Interpretability              High                             Low (requires Explainable
                                                                        AI)
         Computational      Require-   Low                              High
         ments


    • Computational Complexity: These systems demand substantial computational resources,
      which can be a barrier in resource-constrained environments [11].
    • Training Data Requirements : High-quality, labeled datasets are essential for fine-tuning LLMs,
      yet such datasets are often difficult to obtain in cybersecurity domains [14].
    • Interpretability Issues: The black-box nature of LLMs complicates understanding their decision-
      making processes, creating a need for advancements in Explainable AI (XAI) [17].

  To bridge the gap between traditional and LLM-based approaches, researchers are exploring hybrid
models that integrate the efficiency and interpretability of traditional methods with the adaptability
and precision of LLMs. Additionally, ongoing developments in XAI and optimization techniques hold
promise for addressing computational and interpretability challenges, paving the way for broader
adoption of LLM-based IDS in real-world scenarios.


5. Applications in Cybersecurity
The integration of Large Language Models (LLMs) into cybersecurity has enabled significant advance-
ments across various domains. This section explores key applications, highlighting the role of LLMs
in network security, malware analysis, phishing detection, data leakage prevention, and software
vulnerability detection.

5.1. Network Security
LLMs have been instrumental in enhancing network security by detecting anomalies and identifying
potential intrusions. By analyzing log files [14] [7], network traffic [9], and user behaviors [18], LLMs
can:

    • Detect Sophisticated Attacks: Identify advanced persistent threats (APTs) and zero-day vul-
      nerabilities through contextual pattern analysis.
    • Mitigate Distributed Denial of Service (DDoS) Attacks: Recognize traffic anomalies indicative
      of DDoS attacks and trigger real-time mitigation strategies.
    • Enhance Threat Intelligence: Integrate with threat intelligence feeds to provide contextual
      insights into potential vulnerabilities.
                                             Software Vulner-
                                             ability Detection




                   Phishing                         LLMs in                    Network
                   Detection                      Cybersecurity                Security




                                                    Malware
                                                    Analysis


Figure 2: Applications of LLMs in Cybersecurity


5.2. Malware Analysis
Malware detection and classification have greatly benefited from LLMs’ ability to analyze code and
behavioral patterns [19] [20] [21]:
    • Behavioral Analysis: Analyze system logs and executable behaviors to detect malicious activities
      without relying solely on static signatures.
    • Code Analysis: Evaluate and classify obfuscated or polymorphic malware by understanding
      code structures and relationships.
    • Threat Categorization: Facilitate automated categorization of malware families based on
      contextual and structural similarities.

5.3. Phishing and Social Engineering Detection
LLMs have shown exceptional capabilities in detecting phishing attempts and social engineering attacks
[22] [23] [24]:
    • Email and Text Analysis: Identify linguistic patterns and anomalies indicative of phishing
      attempts.
    • Preventing Deceptive Communications: Detect impersonation and spoofing by comparing
      sender identities with behavioral baselines.
    • Proactive Awareness: Generate simulated phishing attempts to train and assess user awareness
      .

5.4. Software Vulnerability Detection
Identifying vulnerabilities in software is a critical application of LLMs in cybersecurity [25] [26] [27]
[28]. By analyzing source code, dependency graphs, and software configurations, LLMs can:
    • Detect Known Vulnerabilities: Automatically identify and flag vulnerabilities from databases
      such as CVE or NVD by recognizing patterns in software code and configurations.
    • Identify Code Smells: Spot potentially risky coding practices that could lead to future vulnera-
      bilities, such as improper input sanitization or outdated dependencies.
    • Enhance Security Testing: Generate test cases to validate software against common exploit
      scenarios, improving the overall robustness of applications.

   These applications demonstrate the versatility of LLMs in addressing complex cybersecurity chal-
lenges, underscoring their value in building proactive and adaptive defense mechanisms.


6. Challenges and Interpretability
The adoption of Large Language Models (LLMs) in Intrusion Detection Systems (IDS) is transformative
but introduces several challenges. Key issues include computational complexity, dataset limitations,
interpretability concerns, ethical considerations, and privacy concerns. Table 2 summarizes these
challenges and proposed solutions.

6.1. Challenges and Proposed Solutions

Table 2
Challenges and Solutions for LLM-Based IDS

           Challenge                     Proposed Solution
           Computational Complexity      Model optimization techniques like quantization and
                                         pruning
           Data Scarcity                 Data augmentation, synthetic data generation, and fed-
                                         erated learning
           Interpretability Issues       Integration of Explainable AI methods such as attention
                                         visualization
           Adversarial Vulnerabilities   Robust loss functions, adversarial training, and ensemble
                                         methods



    • Computational Complexity: LLMs demand high computational resources, presenting barriers
      for real-time detection in resource-constrained environments. Optimization techniques like
      quantization, pruning, and distillation are promising solutions to reduce model size, latency, and
      energy consumption [11].
    • Dataset Limitations: Cybersecurity datasets often suffer from imbalance, scarcity, and privacy
      concerns. These limitations affect model performance, particularly in detecting rare or zero-day
      threats. Techniques such as data augmentation, synthetic data generation, and federated learning
      address these issues while preserving privacy [11].
    • Interpretability Concerns: The ”black-box” nature of LLMs complicates understanding their
      decisions, which is critical for trust and accountability in high-stakes cybersecurity environments.
      Explainable AI (XAI) methods, including attention visualization, feature importance ranking, and
      counterfactual analysis, are being developed to improve transparency [17].
    • Ethical Concerns: LLMs may be vulnerable to adversarial attacks or misuse, such as generating
      phishing emails or automating sophisticated attacks. Countermeasures like adversarial training,
      robust loss functions, and clear policy frameworks are essential to mitigate these risks [29].
    • Privacy Concerns: Training LLMs for cybersecurity often requires sensitive data, which poses
      risks related to data leakage and regulatory non-compliance. Privacy-preserving techniques, such
      as federated learning, differential privacy, and secure multiparty computation, can mitigate these
      risks while allowing the development of effective IDS solutions [29].
Table 3
Emerging Trends in LLM-Based Cybersecurity Solutions

 Trend                        Description
 Real-Time Intrusion De-      The use of LLMs for real-time intrusion detection is gaining traction. Tech-
 tection                      niques such as streaming transformers and lightweight model architectures
                              are being developed to enable real-time processing of network traffic and
                              alerts.
 Federated Learning for       Federated learning frameworks are emerging as a solution to privacy and
 Cybersecurity                data-sharing concerns in cybersecurity. By enabling decentralized model
                              training, federated learning allows organizations to collaborate without ex-
                              posing sensitive data.
 Graph-Based Learning In-     Graph-based models are being combined with LLMs to capture relationships
 tegration                    between entities, such as users, IP addresses, and file hashes. This integration
                              enhances the detection of sophisticated attacks, including supply chain attacks
                              and APTs.
 Explainable AI for Secu-     Explainable AI (XAI) tools are increasingly being adopted to address the inter-
 rity                         pretability challenges of LLMs. Visualization techniques, such as attention
                              maps and saliency scores, are being integrated into cybersecurity workflows
                              to build trust and transparency.


6.2. Future Directions
To overcome these challenges, researchers are focusing on:
   1. Lightweight Model Development: Creating resource-efficient LLMs for edge device deploy-
      ment.
   2. Dataset Expansion: Building diverse, representative datasets to improve training and general-
      ization.
   3. Enhanced XAI: Advancing interpretability to foster trust and transparency in LLM-based
      systems.
   4. Ethical Governance: Establishing guidelines to ensure responsible use and prevent malicious
      exploitation.
By addressing these challenges and adopting innovative solutions, LLM-based IDS can evolve into
scalable, interpretable, and ethically responsible tools for modern cybersecurity.


7. Emerging Trends and Future Directions
LLM advancements, along with AI and cybersecurity progress, are set to revolutionize threat detection
and mitigation. Emerging trends in LLM-based cybersecurity, such as real-time intrusion detection,
federated learning, graph-based learning, and XAI, are highlighted in Table 3. Future improvements,
including energy-efficient models, better datasets, hybrid systems, and enhanced security measures, are
outlined in Table 4.


8. Conclusion
Large Language Models (LLMs) have emerged as powerful tools in transforming cybersecurity, par-
ticularly in enhancing the capabilities of Intrusion Detection Systems (IDS). This paper has reviewed
the methodologies, applications, challenges, and future directions associated with LLM-based IDS,
highlighting their ability to address the limitations of traditional approaches and adapt to evolving
cyber threats.
Table 4
Future Directions for LLM-Based Cybersecurity Solutions

 Direction              Description
 Energy-Efficient       Developing energy-efficient LLMs is crucial for their sustainable deployment in
 LLMs                   cybersecurity. Techniques such as model compression, quantization, and distillation
                        should be further explored to reduce energy consumption.
 Comprehensive          The creation of diverse and representative cybersecurity datasets is essential for
 Cybersecurity          training robust LLMs. Future efforts should focus on curating datasets that capture
 Datasets               a wide range of attack scenarios, including zero-day vulnerabilities and emerging
                        threats.
 Hybrid Models          Hybrid systems that combine traditional IDS techniques with LLM capabilities offer a
                        balanced approach. For instance, combining rule-based detection with LLM-powered
                        anomaly detection can enhance both accuracy and efficiency.
 Adversarial      Ro-   Future research should prioritize developing LLMs that are resilient to adversarial
 bustness               attacks. Adversarial training, ensemble methods, and robust loss functions are
                        promising techniques to enhance model security.
 Ethical and Policy     Establishing clear ethical guidelines and policy frameworks is critical to prevent the
 Frameworks             misuse of LLMs in cybersecurity. Future work should explore mechanisms to enforce
                        responsible deployment and mitigate potential risks.
 Interdisciplinary      Collaboration between AI researchers, cybersecurity professionals, and policymakers
 Collaboration          is essential for driving innovation. Interdisciplinary research can help address complex
                        challenges and unlock new opportunities for LLM-based cybersecurity solutions.


   While LLMs offer significant advantages, such as improved detection accuracy, scalability, and con-
textual analysis, challenges related to computational demands, dataset limitations, and interpretability
persist. Addressing these issues through advancements in model optimization, Explainable AI, and
ethical frameworks will be essential for their sustainable deployment.
   Looking ahead, interdisciplinary collaboration among AI researchers, cybersecurity professionals,
and policymakers will be key to unlocking the full potential of LLMs. By fostering innovation and
addressing critical challenges, LLMs can redefine the cybersecurity landscape, offering scalable and
intelligent solutions for safeguarding digital infrastructures.
   In summary, LLMs represent a promising step toward more adaptive and resilient cybersecurity
systems, paving the way for intelligent, proactive, and scalable defense mechanisms.


Declaration on Generative AI
During the preparation of this work, the authors used ChatGPT, for rephrasing, grammar and spelling
checks, and improving writing style. After using this tool/service, the authors reviewed and edited the
content as needed and takes full responsibility for the publication’s content.


References
 [1] G. Culot, F. Fattori, M. Podrecca, M. Sartor, Addressing industry 4.0 cybersecurity challenges,
     IEEE Engineering Management Review 47 (2019) 79–86.
 [2] D. Hamouda, M. A. Ferrag, N. Benhamida, H. Seridi, Intrusion detection systems for industrial
     internet of things: A survey, in: 2021 International Conference on Theoretical and Applicative
     Aspects of Computer Science (ICTAACS), IEEE, 2021, pp. 1–8.
 [3] M. A. Bouke, A. Abdullah, N. I. Udzir, N. Samian, Overcoming the challenges of data lack,
     leakage, and imensionality in intrusion detection systems: a comprehensive review, Journal of
     Communication and Information Systems 39 (2024).
 [4] A. Vaswani, Attention is all you need, Advances in Neural Information Processing Systems (2017).
 [5] J. Devlin, Bert: Pre-training of deep bidirectional transformers for language understanding, arXiv
     preprint arXiv:1810.04805 (2018).
 [6] T. B. Brown, Language models are few-shot learners, arXiv preprint arXiv:2005.14165 (2020).
 [7] O. G. Lira, A. Marroquin, M. A. To, Harnessing the advanced capabilities of llm for adaptive
     intrusion detection systems, in: International Conference on Advanced Information Networking
     and Applications, Springer, 2024, pp. 453–464.
 [8] M. A. K. Raiaan, M. S. H. Mukta, K. Fatema, N. M. Fahad, S. Sakib, M. M. J. Mim, J. Ahmad, M. E.
     Ali, S. Azam, A review on large language models: Architectures, applications, taxonomies, open
     issues and challenges, IEEE Access (2024).
 [9] M. A. Ferrag, M. Ndhlovu, N. Tihanyi, L. C. Cordeiro, M. Debbah, T. Lestable, N. S. Thandi,
     Revolutionizing cyber threat detection with large language models: A privacy-preserving bert-
     based lightweight model for iot/iiot devices, IEEE Access (2024).
[10] H. Xu, S. Wang, N. Li, K. Wang, Y. Zhao, K. Chen, T. Yu, Y. Liu, H. Wang, Large language models
     for cyber security: A systematic literature review, arXiv preprint arXiv:2405.04760 (2024).
[11] M. A. Ferrag, F. Alwahedi, A. Battah, B. Cherif, A. Mechri, N. Tihanyi, Generative ai and large
     language models for cyber security: All insights you need, arXiv preprint arXiv:2405.12750 (2024).
[12] M. Bayer, P. Kuehn, R. Shanehsaz, C. Reuter, Cysecbert: A domain-adapted language model for
     the cybersecurity domain, ACM Transactions on Privacy and Security 27 (2024) 1–20.
[13] A. Shestov, A. Cheshkov, R. Levichev, R. Mussabayev, P. Zadorozhny, E. Maslov, C. Vadim, E. Buly-
     chev, Finetuning large language models for vulnerability detection, arXiv preprint arXiv:2401.17010
     (2024).
[14] E. Karlsen, X. Luo, N. Zincir-Heywood, M. Heywood, Benchmarking large language models for
     log analysis, security, and interpretation, Journal of Network and Systems Management 32 (2024)
     59.
[15] H. Lai, Intrusion detection technology based on large language models, in: 2023 International
     Conference on Evolutionary Algorithms and Soft Computing Techniques (EASCT), IEEE, 2023, pp.
     1–5.
[16] T. Ali, P. Kostakos, Huntgpt: Integrating machine learning-based anomaly detection and explain-
     able ai with large language models (llms), arXiv preprint arXiv:2309.16021 (2023).
[17] V. Jüttner, M. Grimmer, E. Buchmann, Chatids: Explainable cybersecurity using generative ai,
     arXiv preprint arXiv:2306.14504 (2023).
[18] J. Liu, C. Zhang, J. Qian, M. Ma, S. Qin, C. Bansal, Q. Lin, S. Rajmohan, D. Zhang, Large language
     models can deliver accurate and interpretable time series anomaly detection, arXiv preprint
     arXiv:2405.15370 (2024).
[19] P. M. S. Sánchez, A. H. Celdrán, G. Bovet, G. M. Pérez, Transfer learning in pre-trained large
     language models for malware detection based on system calls, arXiv preprint arXiv:2405.09318
     (2024).
[20] C. Patsakis, F. Casino, N. Lykousas, Assessing llms in malicious code deobfuscation of real-world
     malware campaigns, arXiv preprint arXiv:2404.19715 (2024).
[21] P. Madani, Metamorphic malware evolution: The potential and peril of large language models, in:
     2023 5th IEEE International Conference on Trust, Privacy and Security in Intelligent Systems and
     Applications (TPS-ISA), IEEE, 2023, pp. 74–81.
[22] L. Jiang, Detecting scams using large language models, arXiv preprint arXiv:2402.03147 (2024).
[23] R. Chataut, P. K. Gyawali, Y. Usman, Can ai keep you safe? a study of large language models for
     phishing detection, in: 2024 IEEE 14th Annual Computing and Communication Workshop and
     Conference (CCWC), IEEE, 2024, pp. 0548–0554.
[24] T. Koide, N. Fukushi, H. Nakano, D. Chiba, Detecting phishing sites using chatgpt, arXiv preprint
     arXiv:2306.05816 (2023).
[25] M. A. Ferrag, A. Battah, N. Tihanyi, M. Debbah, T. Lestable, L. C. Cordeiro, Securefalcon: The next
     cyber reasoning system for cyber security, arXiv preprint arXiv:2307.06616 (2023).
[26] M. D. Purba, A. Ghosh, B. J. Radford, B. Chu, Software vulnerability detection using large
     language models, in: 2023 IEEE 34th International Symposium on Software Reliability Engineering
     Workshops (ISSREW), IEEE, 2023, pp. 112–119.
[27] A. Mechri, M. A. Ferrag, M. Debbah, Secureqwen: Leveraging llms for vulnerability detection in
     python codebases, Computers & Security 148 (2025) 104151.
[28] D. Noever, Can large language models find and fix vulnerable software?, arXiv preprint
     arXiv:2308.10345 (2023).
[29] B. C. Das, M. H. Amini, Y. Wu, Security and privacy challenges of large language models: A survey,
     arXiv preprint arXiv:2402.00888 (2024).