<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Neurosymbolic AI in Digital Forensics: Commonsense and Qualitative Reasoning</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Alessia Donata Camarda</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Calabria</institution>
          ,
          <addr-line>Rende, 87036</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>Technological devices have become an integral part of our daily lives. Although apparently harmless, many of these contain a huge amount of information about us which can be relevant to incriminate the culprit of a crime. Due to this reason, it is essential to include actions concerning the collection and analysis of digital evidence in the several phases of the investigative process: this is where Digital Forensics was born. Despite the current fame gained by the field, the context in which it is placed requires particular attention in the implementation of frameworks and methods aligned with principles such as transparency, accountability, and fairness. My research proposal aims to leverage new Neurosymbolic Artificial Intelligence approaches to build tools and explore the possibility of automating tasks in Digital Forensics. Traditional tools alone are currently not enough to provide valid and concrete help to the field: it is thus necessary to coordinate the use of newer methods that are increasingly present in the panorama of Artificial Intelligence and Automation to tackle new tasks or re-explore already seen ones, but from a Trustworthy perspective. The main ingredients useful to accomplish this task will be Commonsense and Qualitative reasoning, Answer Set Programming, and Large Language Models.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Answer Set Programming</kwd>
        <kwd>Neurosymbolic AI</kwd>
        <kwd>Digital Forensics</kwd>
        <kwd>Commonsense reasoning</kwd>
        <kwd>Qualitative Reasoning</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Traditional forensics, i.e. the set of activities carried out during the life cycle of an investigation,
has been an omnipresent field directly related to the presence and activities of humans. Discovering
actions, decisions, or intent of specific individuals and disambiguating identities are among its main
purposes, which is why removing the human component from such processes is almost impossible.
Before the advent of computers, forensics primarily dealt with physical evidence, such as fingerprints,
blood, or handwritten documents. However, with the spread of technology, criminal activity also
changed, increasingly involving the use of computers, thus making traditional forensic techniques
insuficient. To address these new forms of evidence (e.g., files, logs, and so on), the investigation process
has become digital, leading to the rise of Digital Forensics (DF) [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ] . DF involves the identification,
collection and analysis of digital evidence, i.e., all the information extracted and obtained using electronic
instruments. The rapid development of this field has entailed the birth of a wide variety of subfields,
such as Computer Forensics, Network Forensics, IoT Forensics, Incident Response, and so on. All of these
deal with diferent aspects related to the use of computers and other technologies within criminal
cases. Digital forensics is part of an even larger branch of computer science: Cybersecurity [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], which
concerns the protection of digital systems, data, and services from malicious activities carried out by
attackers. Digital forensics thus integrates knowledge and methodologies from a variety of disciplines,
although it remains primarily rooted in computer science. From its birth, DF has been object of
study for researchers, and new scientific discoveries have gradually driven the progress of this field.
Consequently, the spread of Artificial Intelligence (AI) has also significantly influenced the development
of tools supporting the investigative process [
        <xref ref-type="bibr" rid="ref4 ref5 ref6">4, 5, 6</xref>
        ]. Clear examples of how AI has become integral
to this field include tools for analyzing electronic devices and biometric identification systems that
combine facial recognition, fingerprints, iris scans, and more. Despite their usefulness, such systems are
often the focus of strong criticism, especially with regard to the privacy and security of the individuals
involved [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Consequently, the call for Trustworthy AI (TAI), which emphasizes privacy and ethics, has
gained prominence in response to these concerns. TAI aims to develop AI systems whose core values are
ethics, safety, transparency, and human rights and values. Several institutional entities have proposed
regulations and guidelines to be followed by AI systems. For example, the European Commission [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]
has established 7 key requirements that Trustworthy AI is expected to meet. Despite these guidelines,
it is not always possible to fully adhere to them. For example, deep learning-based systems cannot
easily ensure transparency in their operations. Moreover, several cases have demonstrated that an
improper use of such models can even result in guidelines violation [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. This demonstrates that there
are still major steps to be taken to develop systems that can fully comply with the guidelines while still
remaining capable of performing complex tasks.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Related work</title>
      <p>
        Several eforts have been made trying to provide meaningful contributions to the field. One of the key
principles on which research is focusing is explainability. In [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], a first formal definition of Explainable
Artificial Intelligence in Cyber Security is proposed. The authors also ofer a detailed study about which
system properties are necessary to ensure AI-powered tools aligned with legal and ethical standards.
Thanks to the spread of deep learning, many tasks that until now were only performed by humans have
started to be (even if partially) solved. This is also true for digital forensics and cybersecurity, which
has seen a great contribution from this community. For example, in [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] it is introduced ForensicLLM, a
Large Language Model fine-tuned on Question and Answering samples extracted from digital forensic
literature. This model is designed to support experts in their daily work by helping them analyze
forensic artifacts, answer technical questions, or suggest investigative directions. In [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], an holistic
and generalized framework for DF is proposed for standardization. Many of these attempts try to
integrate explainability techniques within the deep learning pipeline [
        <xref ref-type="bibr" rid="ref12 ref13">12, 13</xref>
        ]. Recent advances have
also leveraged Neuro-Symbolic approaches to enhance explainability. For instance, [14] proposes a
method to extract logic-based global explanations from convolutional neural networks, supporting
interpretability in tasks like image classification. Eforts have not only focused on Deep learning and
Machine learning, but some approaches that exploit logical formalisms have also been proposed. In
this regard, we mention DigForASP [15], a COST Action that aims to create a cooperation network for
exploring the potential of the application of logic-based Artificial Intelligence in the Digital Forensics
ifeld. This cooperation has produced several results [ 16, 17, 18]. Similarly, in [19, 20], attempts are made
to formalize knowledge related to criminal investigations. These works aim to develop a structured
vocabulary capable of describing investigations activities, events and digital artifacts at diferent levels
of detail and according to the task to be performed.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Goal of the research</title>
      <p>The main goal of my research is to leverage Neurosymbolic artificial intelligence [ 21] to bridge the current
gaps in digital forensics. Since traditional tools alone are currently insuficient to provide efective and
concrete support to the field, it remains necessary to exploit deep learning and, in particular, Large
Language Models [22] to perform support tasks. On the other hand, it is necessary to adhere as closely
as possible to the Guidelines for a Trustworthy AI. In this respect, logical formalisms, such as Answer
Set Programming (ASP) [23], allow to enforce transparency and explainability while simultaneously
enabling reasoning over incomplete or conflicting evidence. To address two of the most critical aspects
of digital forensics - namely, the world knowledge possessed by humans and the partial, uncertain
information available during crime investigations - we propose integrating commonsense and qualitative
reasoning into our pipelines:
Commonsense reasoning. Commonsense understanding about daily life, typical movement patterns
for normal or abnormal behaviors, and specific expertise related to investigative actions can be useful
for identifying contradictions between depositions and the evidence available in a case. Therefore,
such knowledge must be included among the information relevant for resolving the case. Afordance
knowledge, i.e., information about what can or cannot be done with an object, can be mixed with
case-specific knowledge. This can help in understanding, e.g., how certain objects are linked to the
actions performed by the agents involved. Taking this type of information into account is important
because humans acquire such knowledge over time and take it for granted, whereas computer systems
do not possess it by default. It is necessary to find ways to enable them to learn it. On this matter, Answer
Set Programming is particularly well-suited to represent new knowledge and exceptions, enabling the
construction of a declarative knowledge base that encodes commonsense and afordance reasoning and
supports efective inference over it.</p>
      <p>Qualitative reasoning. The data available in a given case are not always complete or quantitatively
precise. Indeed, law enforcement and investigators often have to work with incomplete and insuficient
data. Moreover, individuals with expertise in computer science and related fields may attempt to
tamper with digital evidence to hide the truth, further compromising the reliability of the available
data. To implement reasoning and analysis processes, it is therefore necessary to leverage Knowledge
Representation and Reasoning tools such as qualitative reasoning, which allows reasoning about
qualitative notions, e.g., near, hotter than. Qualitative reasoning can be combined with rule-based
systems or relational formalisms (e.g., temporal or spatial logic frameworks) to interpret incomplete data,
recognize patterns, reconstruct scenarios, perform behavioral analysis, and automate the exploration of
investigative hypotheses. On this matter, Answer Set Programming is particularly well-suited, as it
enables the representation of incomplete or uncertain knowledge and supports inference over symbolic
relations, even in the absence of reliable or quantitative data.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Current status of the research and preliminary results</title>
      <p>Ongoing research is addressing the aforementioned topics. The following section presents some
preliminary considerations and findings obtained so far.</p>
      <p>Contradictions management. Contradiction management has been object of study in many fields
under a diferent terminology. Since forensics field deals with natural language, and therefore with
its complexities, e.g. ambiguity, traditional tools cannot provide useful results. Several approaches
have attempted to use Deep Learning to detect contradictions [24, 25], even in the legal field [ 26].
Although the results are good, they are still far from optimal. We started to study the application of
Neuro-Symbolic AI to contradiction management. In this context, we propose a pipeline that relegate
the role of Large Language Models to support tasks, such as commonsense knowledge extraction and
translation of input sentences into structured format. Then, the reasoning phase is performed by an
Answer Set Programming solver, thus allowing us to justify the gained conclusions. We managed to
obtain about 84% accuracy on the dataset at hand. The results are presented in the following paper
[27].</p>
      <p>Temporal reasoning. Time management and reasoning about events are among the fundamental
issues in digital forensics. When dealing with imprecise information, the results are afected by this
uncertainty, which also applies to unreliable timestamps and overlapping events, thereby significantly
influencing the outcome of an investigation. Diferent approaches have been proposed to address
temporal reasoning under uncertainty, including Allen’s interval algebra [28], but these methods are
often dificult to integrate into automated reasoning systems. We are currently developing a system
that supports temporal reasoning by approximating the available values and considering multiple
scenarios simultaneously, while accounting uncertainty and managing it. Always keeping in mind
the Trustworthy AI guidelines, our goal is to propose an explainable system capable of clarifying the
reasoning behind the obtained results, despite the uncertainty of the available information.
Commonsense extraction and anomaly detection. Neuro-Symbolic AI is a discipline that can
provide valuable support in various fields and has become the focus of numerous studied. For example, it
can be particularly useful in areas such as anomaly detection and commonsense reasoning. In particular,
we are investigating how to detect anomalies that contradict commonsense or default assumptions,
rather than identifying outliers in numerical data. The pipeline under consideration leverages LLMs to
extract commonsense knowledge about objects, people, and their typical attributes. Similar work has
been carried out with the aim of populating ontologies using LLMs [29]. In our proposed architecture,
the reasoning phase is handled by an Answer Set Programming solver, where a set of rules detects
unexpected simple objects configurations or anomalous action executions.</p>
      <sec id="sec-4-1">
        <title>Decision Support Framework for Trustworthy AI. Digital forensics operates across multiple</title>
        <p>contexts, each posing unique challenges, which makes the demand for trustworthy AI one of the
most pressing. In this context, we propose the Socio-Technical framework for T rustworthy Artificial
I ntelligence in Digital Forensics, STeForTAI, a theoretical and methodological framework. Its conception
and design emerged from several meeting of the DigForASP project. STeForTAI is grounded in
SocioTechnical Systems Theory, which support the analysis of complex systems involving both technical and
social components, and aligns with the European Commission’s Ethics Guidelines for Trustworthy AI.
The framework was formally introduced in [30] and is currently undergoing peer review.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Open issues and expected achievements</title>
      <p>Given its inherently human-centered nature, some of the open issues in digital forensics relate to its
interaction with the people involved in the use of forensic tools. In addition, the increasingly involvement
of artificial intelligence in the development of such tools brings with it-self several technological and
methodological challenges. Below, we outline some of the most important open issues in this field and
suggest some possible contributions.</p>
      <p>Privacy concerns and Accountability. Criminal cases involve actions performed by individuals;
thus, all the evidence collected relates to what these people did and where they were at specific points in
time. Although this information is necessary for conducting an investigation, it raises privacy concerns
regarding what information is collected, how it is used and who can access it. Furthermore, much
of this knowledge, even if accessed during an investigation, cannot be retained by individuals who
lack authorization. However, such data would be valuable for training models or developing tools
to analyze it. This results in a significant lack of datasets for researchers, who can therefore seldom
test their tools on real data. In addition, this prevents the standardization of automation and analysis
processes, thus hindering the reproducibility and validation of proposed methods. In this regard, this
research is particularly focused on building a forensic commonsense knowledge dataset based on already
existing commonsense datasets. Furthermore, anonymization techniques could enable the creation of
anonymized datasets that would support the initiation of a benchmark standardization process.</p>
      <sec id="sec-5-1">
        <title>The complex task of transforming available knowledge into an useful format. The evidence</title>
        <p>collected may include data extracted from private digital devices, surveillance cameras, or physical
elements found at the crime scene. To represent complex systems, such as the environments in which
we live and interact, it is necessary to create simplified representations that can be more easily managed.
However, determining the appropriate level of abstraction is not always easy: a higher level of abstraction
can lead to the loss of important information, whereas a lower level can result in retaining details that
complicate rather than clarify the problem. In this regard, various information extraction techniques,
such as entity and relations extraction, could be explored with the support of LLMs. The resulting
knowledge can then be represented using Answer Set Programming.</p>
        <p>Lack of explainability. Answers provided by deep learning models are often not explainable, or
the explanations produced are dificult for non-experts to understand. This raises the issue of how
to explain to non-experts what was done and the process that led to specific results. Explainability
is essential in fields such as forensics, where both the results and the processes used to obtain them
must be clear, especially to those involved (e.g., judges who must issue sentences or individuals whose
lives depend on such decisions). In this respect, since this research delegates the reasoning component
to explainable solvers whose results are fully deterministic and reproducible, it could explore the
application of neurosymbolic techniques both for unsolved tasks and for solved tasks whose existing
methods lack explainability.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>This work was partially supported by project SERICS (PE00000014) under the MUR National Recovery
and Resilience Plan funded by the European Union - NextGenerationEU.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>The author used GPT-4o to do grammar and spelling check and rewriting of specific sentences. After
using these tools, the authors reviewed and edited the content as needed and take full responsibility for
the publication’s content.
[14] P. Padalkar, H. Wang, G. Gupta, Nesyfold: A framework for interpretable image classification, in:</p>
      <p>AAAI, AAAI Press, 2024, pp. 4378–4387.
[15] S. Costantini, F. A. Lisi, R. Olivieri, Digforasp: A european cooperation network for logic-based AI
in digital forensics, in: CILC, volume 2396 of CEUR Workshop Proceedings, CEUR-WS.org, 2019, pp.
138–146.
[16] F. A. Lisi, G. Sterlicchio, Mining sequences in phone recordings with answer set programming, in:
HYDRA/RCRA@LPNMR, volume 3281 of CEUR Workshop Proceedings, CEUR-WS.org, 2022, pp.
34–50.
[17] F. A. Lisi, G. Sterlicchio, Declarative AI and digital forensics: Activities and results within the
digforasp project, in: Ital-IA, volume 3486 of CEUR Workshop Proceedings, CEUR-WS.org, 2023, pp.
437–442.
[18] J. Medina-Moreno, Digital forensics: Evidence analysis via intelligent systems and practices
digforasp - CA17124. challenges and achievements: Plenary talk, in: SISY, IEEE, 2022, pp. 17–18.
[19] S. Costantini, G. D. Gasperis, R. Olivieri, Digital forensics and investigations meet artificial
intelligence, Ann. Math. Artif. Intell. 86 (2019) 193–229.
[20] D. Kahvedzic, M. T. Kechadi, DIALOG: A framework for modeling, analysis and reuse of digital
forensic knowledge, Digit. Investig. 6 (2009) S23–S33.
[21] M. K. Sarker, L. Zhou, A. Eberhart, P. Hitzler, Neuro-symbolic artificial intelligence: Current
trends, CoRR abs/2105.05330 (2021).
[22] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, I. Polosukhin,</p>
      <p>Attention is all you need, in: NIPS, 2017, pp. 5998–6008.
[23] V. Lifschitz, Answer Set Programming, Springer, 2019.
[24] V. Lingam, S. Bhuria, M. Nair, D. Gurpreetsingh, A. Goyal, A. Sureka, Deep learning for conflicting
statements detection in text, PeerJ Prepr. 6 (2018) e26589.
[25] M. de Marnefe, A. N. Raferty, C. D. Manning, Finding Contradictions in text, in: ACL, The</p>
      <p>Association for Computer Linguistics, 2008, pp. 1039–1047.
[26] S. Surana, S. Dembla, P. Bihani, Identifying Contradictions in the Legal Proceedings Using Natural</p>
      <p>Language Models, SN Comput. Sci. 3 (2022) 187.
[27] A. D. Camarda, G. Ianni, A study on contradiction detection using a neuro-symbolic approach, in:</p>
      <p>CILC, volume 4003 of CEUR Workshop Proceedings, CEUR-WS.org, 2025.
[28] M. Grüninger, Z. Li, The time ontology of allen’s interval algebra, in: TIME, volume 90 of LIPIcs,</p>
      <p>Schloss Dagstuhl - Leibniz-Zentrum für Informatik, 2017, pp. 16:1–16:16.
[29] G. Ciatto, A. Agiollo, M. Magnini, A. Omicini, Large language models as oracles for instantiating
ontologies with domain-specific knowledge, Knowl. Based Syst. 310 (2025) 112940.
[30] A. Brännström, A. D. Camarda, S. Costantini, P. Dell’Acqua, C. Gallese, G. Ianni, F. A. Lisi,
V. Mascardi, J. C. Nieves, Supporting trustworthiness in socio-technical frameworks with logic
programming, 2024. Submitted.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Pollitt</surname>
          </string-name>
          ,
          <article-title>A history of digital forensics</article-title>
          ,
          <source>in: IFIP Int. Conf. Digital Forensics</source>
          , volume
          <volume>337</volume>
          <source>of IFIP Advances in Information and Communication Technology</source>
          , Springer,
          <year>2010</year>
          , pp.
          <fpage>3</fpage>
          -
          <lpage>15</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>R.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <article-title>Digital evidence and computer crime: Forensic science, computers and the internet</article-title>
          ,
          <source>Int. J. Law Inf. Technol</source>
          .
          <volume>11</volume>
          (
          <year>2003</year>
          )
          <fpage>98</fpage>
          -
          <lpage>100</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Alam</surname>
          </string-name>
          , Cybersecurity: Past, present and future,
          <year>2024</year>
          . URL: https://arxiv.org/abs/2207.01227. arXiv:
          <volume>2207</volume>
          .
          <fpage>01227</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Khalid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Iqbal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. C. M.</given-names>
            <surname>Fung</surname>
          </string-name>
          ,
          <article-title>Towards a unified xai-based framework for digital forensic investigations</article-title>
          ,
          <source>Digit. Investig</source>
          .
          <volume>50</volume>
          (
          <year>2024</year>
          )
          <fpage>301806</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A.</given-names>
            <surname>Wickramasekara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Breitinger</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Scanlon, Exploring the potential of large language models for improving digital forensic investigation eficiency</article-title>
          ,
          <source>Forensic Sci. Int. Digit. Investig</source>
          .
          <volume>52</volume>
          (
          <year>2025</year>
          )
          <fpage>301859</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Solanke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Biasiotti</surname>
          </string-name>
          ,
          <article-title>Digital forensics AI: evaluating, standardizing and optimizing digital evidence mining techniques</article-title>
          ,
          <source>Künstliche Intell</source>
          .
          <volume>36</volume>
          (
          <year>2022</year>
          )
          <fpage>143</fpage>
          -
          <lpage>161</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>W. K.</given-names>
            <surname>Jung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. Y.</given-names>
            <surname>Kwon</surname>
          </string-name>
          ,
          <article-title>Privacy and data protection regulations for AI using publicly available data: Clearview AI case</article-title>
          ,
          <source>in: ICEGOV, ACM</source>
          ,
          <year>2024</year>
          , pp.
          <fpage>48</fpage>
          -
          <lpage>55</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>European</given-names>
            <surname>Commission</surname>
          </string-name>
          ,
          <article-title>Ethics guidelines for trustworthy AI</article-title>
          , https://ec.europa.
          <article-title>eu/ digital-single-market/en/news/ethics-guidelines-trustworthy-</article-title>
          <string-name>
            <surname>ai</surname>
          </string-name>
          ,
          <year>2019</year>
          . [Online].
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J.</given-names>
            <surname>Angwin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Larson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mattu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Kirchner</surname>
          </string-name>
          ,
          <article-title>Machine bias: There's software used across the country to predict future criminals. and it's biased against blacks</article-title>
          ,
          <source>ProPublica</source>
          (
          <year>2016</year>
          ). [Online].
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>S.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Altiparmak</surname>
          </string-name>
          ,
          <article-title>Xai-cf - examining the role of explainable artificial intelligence in cyber forensics</article-title>
          ,
          <year>2024</year>
          . URL: https://arxiv.org/abs/2402.02452. arXiv:
          <volume>2402</volume>
          .
          <fpage>02452</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>B.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ghawaly</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>McCleary</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Webb</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Baggili</surname>
          </string-name>
          ,
          <article-title>Forensicllm: A local large language model for digital forensics</article-title>
          ,
          <source>Digit. Investig</source>
          .
          <volume>52</volume>
          (
          <year>2025</year>
          )
          <fpage>301872</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>M.</given-names>
            <surname>Zolanvari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <surname>K. M. Khan</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Jain</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Meskin</surname>
          </string-name>
          , TRUST XAI:
          <article-title>model-agnostic explanations for AI with a case study on iiot security</article-title>
          ,
          <source>IEEE Internet Things J</source>
          .
          <volume>10</volume>
          (
          <year>2023</year>
          )
          <fpage>2967</fpage>
          -
          <lpage>2978</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>M.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>An explainable machine learning framework for intrusion detection systems</article-title>
          ,
          <source>IEEE Access 8</source>
          (
          <year>2020</year>
          )
          <fpage>73127</fpage>
          -
          <lpage>73141</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>