<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Logically Explainable Malware Detection</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Peter Anthony</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Francesco Giannini</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Michelangelo Diligenti</string-name>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marco Gori</string-name>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Martin Homola</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Stefan Balogh</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ján Mojžiš</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Comenius University Bratislava</institution>
          ,
          <country country="SK">Slovakia</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Scuola Normale Superiore</institution>
          ,
          <addr-line>Pisa</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Slovak Academy of Sciences Bratislava</institution>
          ,
          <country country="SK">Slovakia</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Slovak Technical University Bratislava</institution>
          ,
          <addr-line>slovakia</addr-line>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>University of Siena</institution>
          ,
          <addr-line>Siena</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Malware detection is a challenging application due to the rapid evolution of attack techniques, and traditional signature-based approaches struggle with the high volume of malware samples. Machine learning approaches face such limitation, but lack a clear interpretability, whereas interpretable models often underperform. This paper proposes to use Logic Explained Networks (LENs), a recently proposed class of interpretable neural networks that provide explanations using First-Order Logic rules, for malware detection. Applied to the EMBER dataset, LENs show robustness superior to traditional interpretable methods and performance comparable to black-box models. Additionally, we introduce a tailored LEN version improving the fidelity of logic-based explanations.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Malware Detection</kwd>
        <kwd>Explainable AI</kwd>
        <kwd>First-Order Logic</kwd>
        <kwd>Logic Explained Networks</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Malware detection is crucial in cybersecurity due to the rapid evolution of attack techniques, and
traditional signature-based methods from companies like Comodo, Kaspersky, and Symantec
struggle to keep up with the millions of new malware samples each year [
        <xref ref-type="bibr" rid="ref1 ref2 ref3">1, 2, 3</xref>
        ]. Machine
learning, particularly Deep Neural Networks (DNN), ofers robust solutions by recognizing
complex patterns, handling large datasets, and detecting zero-day attacks. However, these
methods often lack explainability, limiting their trustworthiness in safety-critical applications.
      </p>
      <p>
        Recently, Logic Explained Networks (LENs) [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] have been proposed as an
explainable-bydesign class of neural networks. LENs use human-understandable predicates and provide
explanations through First-Order Logic (FOL) formulas, balancing accuracy and interpretability.
Although LENs have shown success in various domains [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] and to diferent kind of data, such as
images [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], textual information [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] and graphs [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], their efectiveness on large datasets like the
EMBER malware dataset [9] (800,000 samples with thousands of features) remains unexplored.
This paper demonstrates that LENs can form a robust malware detection framework with
competitive performance against black-box models and superior to other interpretable methods.
Additionally, we introduce an innovative approach to enhance the fidelity of LENs’ explanations,
making them more accurate and meaningful.
      </p>
      <p>This paper makes three main contributions: (i) it shows that Logic Explained Networks are
efective for malware detection, providing meaningful explanations and predictive performance
comparable to state-of-the-art black box models while outperforming other interpretable models,
(ii) it introduces an improved rule extraction process for LENs, enhancing scalability, fidelity,
complexity, and predictive accuracy, and (iii) it ofers an in-depth analysis of the extracted rules,
evaluating their fidelity, complexity, and accuracy as input feature size increases.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>Machine Learning Approaches. Machine learning techniques are commonly used to train
malware detectors and uncover complex patterns in malicious software [10, 11, 12]. While deep
learning shows promising results due to its ability to learn from large datasets and generalize to
unknown samples [13], several limitations remain, including low generalization performance
for unseen malware samples. Moreover, classical machine learning models act as black-boxes,
hindering explainability. In security-critical domains, interpretability is crucial for trust and legal
compliance [14, 15, 16, 17]. Indeed, cybersecurity experts need insights into model decisions to
enhance system trustworthiness.</p>
      <p>
        Interpretable AI Models. Interpretable models (e.g., linear regression, decision trees) ofer
explanations, but struggle with complex features [18]. These methods prioritize
interpretability over performance, unlike deep neural networks [19]. To address this trade-of, various
techniques have been proposed. One such method is permutation feature importance, which
interprets a wide range of machine learning models but comes with high computational costs [20].
Additionally, surrogate model methods like LIME [21] and SHAP [22] approximate the target
model using interpretable models. However, their expressive ability may not match that of the
complex target model, leading to inaccurate interpretations [
        <xref ref-type="bibr" rid="ref9">23</xref>
        ]. For the malware detection
task, Švec et al. [
        <xref ref-type="bibr" rid="ref10">24</xref>
        ] explored interpretable concept learning algorithms all implemented in
DL-Learner: OCEL (OWL Class Expression Learner), CELOE (Class Expression Learning for
Ontology Engineering), PARCEL (Parallel Class Expression Learner), and SPARCEL
(Symmetric Parallel Class Expression Learner). Their approach provided clear explanations but faced
computational challenges and low performance.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Background on Logic Explained Networks</title>
      <p>
        Logic Explained Networks [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] combine the advantages of black-boxes and transparent models
by providing promptly interpretable neural networks in First-Order Logic (FOL). LENs take
human-understandable predicates as inputs, such as tabular data or concepts extracted from raw
data, and express explanations in FOL rules involving these predicates. Thanks to their complex
neural processing, LENs achieve high-level performance while being easily interpretable.
(a) Local explanation for single sample
(b) Class-level explanation
      </p>
      <p>
        Formally, a LEN  can be defined as a map from  = [
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ]-valued input concepts to
 = [
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ] output classes, which can be used to directly classify samples and provide meaningful
local and/or global explanations (cf. Figure 1).
      </p>
      <p>
        As a special case, in malware detection we have  = 2 classes, i.e. {malware, benign}. In
general, for each sample , with  ∈ {1, . . . , }, a prediction () = 1 is locally explained
by the conjunction of the most relevant input features () = ⋀︀∈()(¬) , where  is
a logic predicate associated with the -th input feature, and () is the set of relevant input
features for the -th task. Notice that each  can occur as a positive  or negative ¬ literal,
according to a given threshold (e.g. 0.5). The most representative local explanations can be
aggregated to get a global explanation  = ⋁︀()∈() (), where () collects the -most
frequent local explanations for the class  in the training set. We will refer later to such global
explanations as raw LEN explanations. To prevent too complex global explanations, Gabriele
et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] suggests a top-(k) strategy, which focuses on aggregating only the most accurate local
explanations. This method, we will refer to as standard LEN explanations, only includes local
explanations that contribute to an improvement in validation accuracy, thus ensuring that the
generalization of the rules is efective when applied to new data sets. However, even with this
strategy, the complexity can remain high for large datasets with many samples and features.
      </p>
    </sec>
    <sec id="sec-4">
      <title>4. Tailored-LENs’ Explanations</title>
      <p>
        Standard LEN explanations are more accessible than raw ones but still have some drawbacks,
such as (i) determining the optimal -value can be computationally intensive, (ii) selecting
top- local explanations based on their individual accuracy tends to select explanations that
increase false positives rate, due to favoring high recall over precision, which is not preferable
to construct a robust discrimination against malware. To enhance global LEN explanations
for malware detection, this paper introduces what we call the Tailored-LEN explanation
method, which uses a line search optimization to find the best threshold for choosing the
right combination of local explanations, and removing terms from outlier samples to improve
explanation quality. More specifically, we used a precision threshold to aggregate the local
explanations, aiming to reduce false positives and avoid misrepresenting the model. Then,
Model
LGBM[
        <xref ref-type="bibr" rid="ref13">27</xref>
        ]
ANN/DNN[
        <xref ref-type="bibr" rid="ref13">27</xref>
        ]
Improved DNN[
        <xref ref-type="bibr" rid="ref14">28</xref>
        ]
FFN[
        <xref ref-type="bibr" rid="ref15">29</xref>
        ]
CNN[
        <xref ref-type="bibr" rid="ref15">29</xref>
        ]
MalConv w/ GCG[
        <xref ref-type="bibr" rid="ref16">30</xref>
        ]
10
100
1000
2000
_
      </p>
      <p>No
No
No
No
No
No
Yes
Yes
Yes
Yes
Yes
an optimization process iteratively adjusts this threshold to find an optimal balance between
precision and recall. The best solution is a simplified formula that improves validation accuracy,
and only beneficial local explanations are included in the final Tailored-LEN global explanation.
The details of this method can be found in Algorithm 1 in the Appendix.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Experiments</title>
      <p>
        All the experiments carried out in this section are based on the EMBER dataset [9], a well-known
dataset for malware detection with 800, 000 labelled (400, 000 benign and 400, 000 malicious)
and 300, 000 unlabelled samples, respectively. For our experimental analysis we utilized the
version with derived features, as defined by Mojžiš and Kenyeres [
        <xref ref-type="bibr" rid="ref11">25</xref>
        ], which represent a
variation of the ontology realized by Švec et al. [
        <xref ref-type="bibr" rid="ref12">26</xref>
        ]. More details about the experimental
settings can be found in Appendix A.1. The experiments aim to demonstrate that: (i) LENs
perform similarly to black-box models while ofering explanations; (ii) LENs surpass previously
used interpretable machine learning models; (iii) Tailored-LENs explanations provide a better
trade-of in terms of complexity vs accuracy vs fidelity wrt standard and raw LEN explanations.
      </p>
      <p>Comparison against black-box models. We compare LENs with black-box models using
the full EMBER dataset, with 600k samples allocated for training and 200k samples for testing. We
also evaluated the performances of LENs varying diferent subsets’ size of the most informative
features, ranging from 10 to 2000 features, to observe the impact on the model’s results.
Results. Table 1 shows that LENs have high performances in malware detection, achieving
an accuracy of at least 92.07% and an F1-score of 92.17% with a minimum of 100 features. The
performances are slightly lower with the full feature set, possibly due to the feature binarization
process, which increases the feature dimensionality and potentially introduces noise. Despite
this, LENs closely match the best deep-learning black-box models, with less than a 5% diference
in most metrics. Notably, LENs are competitive even with a small feature set and maintain high
generalization capabilities with larger feature sizes. In addition, the key advantage of LENs over
black-box models is their interpretability, which provides insights into the decision-making
process.</p>
      <p>
        Comparison against interpretable approaches. We identified two significant
contributions for interpretable malware detection using the EMBER dataset: (i) the concept learning
method by Švec et al. [
        <xref ref-type="bibr" rid="ref10">24</xref>
        ], and (ii) the decision-tree-based techniques by Mojžiš and
Kenyeres [
        <xref ref-type="bibr" rid="ref11">25</xref>
        ]. The former approach, due to its complexity, was tested on 5,000 random samples
with 5-fold cross-validation. The latter utilized a dataset of 600,000 samples with an 80%/20%
training/testing for 25, 50, 75, 100 and 200 feature sets. To ensure consistency, the same sample
size and cross-validation method were used for the concept learning approach, and the same
feature set and sample distribution were applied for the decision-tree-based approach.
Results. LENs significantly outperform concept learning approaches in all metrics (cf. Table 2).
This support the claim that LENs can represent a promising solution for real-world deployment,
meeting the increasing need for both clarity and efectiveness in malware detection systems.
Additionally, when compared with standard decision tree models, LENs demonstrate good
performance wrt diferent feature sizes, as evidenced in Figure 2. While LENs outperform most
Accuracies over Different Feature Sizes
      </p>
      <p>LEN Raw Explanation
0.65 TSatailonrdeadr-dL-ELNENExEpxlpl</p>
      <p>LEN
0.60 5.0 7.5 10.0 12.5 15.0 17.5 20.0 22.5 25.0</p>
      <p>Feature Size
(a) Comparison based on accuracy
1.0 1.0</p>
      <p>Fidelities over different Feature Sizes
1.0 0.98 0.96
decision tree models, their performance is on par with the J48-ESFS model.</p>
      <p>
        Analysis of provided Explanations. The Tailored-LEN explanation method (Section 4) was
evaluated against the Raw-LEN and Standard-LEN methods proposed in the original paper [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
This evaluation used 25,000 samples and tested feature sets of 5, 10, 15, 20, and 25 features. with
the data divided into a 75% and a 25% split for training and testing, respectively.
Results. Figure 3 illustrates that Tailored LEN explanations outperform Standard LENs
across all feature sizes, ofering better fidelity and lower complexity. Raw LENs provide a
better predictive performance but sufer from high complexity, making them less interpretable.
Standard-LENs, while similar in complexity to Tailored-LENs, fall short in both fidelity and
accuracy compared to Tailored-LENs. Additionally, the practicality of the extracted rules was
analyzed in the context of malware detection applications by a cybersecurity expert (cf. Table 3
in Appendix).
      </p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusions</title>
      <p>This paper studies the application of Logic Explained Networks to malware detection. The
conducted experiments demonstrate that LENs can attain performance comparable to complex
black-box neural models, while maintaining explenability and outperforming other interpretable
machine learning alternatives in terms of eficacy.</p>
      <p>Furthermore, this study introduces a novel algorithm designed to extract global explanations
from LENs. This algorithm improves the predictive precision of LENs, while yielding
explanations characterized by both elevated fidelity and reduced complexity. The findings support the
claim that LENs are a promising candidate for the integration of explainable methodologies
into malware detectors.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>This work was supported by the TAILOR Collaboration Exchange Fund (CEF) under the
European Union’s Horizon 2020 research and innovation program GA No 952215. This work
has been also supported by the Partnership Extended PE00000013 - “FAIR - Future Artificial
Intelligence Research” - Spoke 1 “Human-centered AI”. The work was also sponsored by the
Slovak Republic under the grant no. APVV-19-0220 (ORBIS).
[9] H. S. Anderson, P. Roth, Ember: an open dataset for training static pe malware machine
learning models, arXiv preprint arXiv:1804.04637 (2018).
[10] S. Millar, N. McLaughlin, J. Martinez del Rincon, P. Miller, Z. Zhao, Dandroid: A
multiview discriminative adversarial network for obfuscated android malware detection, in:
Proceedings of the tenth ACM conference on data and application security and privacy,
2020, pp. 353–364.
[11] S. Millar, N. McLaughlin, J. M. del Rincon, P. Miller, Multi-view deep learning for
zeroday android malware detection, Journal of Information Security and Applications 58
(2021) 102718. URL: https://www.sciencedirect.com/science/article/pii/S2214212620308577.
doi:https://doi.org/10.1016/j.jisa.2020.102718.
[12] Y. Ye, L. Chen, S. Hou, W. Hardy, X. Li, Deepam: a heterogeneous deep learning framework
for intelligent malware detection, Knowledge and Information Systems 54 (2018) 265–285.
[13] P. S. H. K. J. Hirpara, Exploring the diverse applications of deep learning across multiple
domains, Recent Research Reviews Journal 4 (2023) 100692. URL: https://irojournals.com/
rrrj/articles/view/2/1/16. doi:http://dx.doi.org/10.36548/rrrj.2023.1.16.
[14] G. Iadarola, F. Martinelli, F. Mercaldo, A. Santone, Towards an interpretable deep learning
model for mobile malware detection and family identification, Computers &amp; Security 105
(2021) 102198. URL: https://www.sciencedirect.com/science/article/pii/S0167404821000225.
doi:https://doi.org/10.1016/j.cose.2021.102198.
[15] A. Mills, T. Spyridopoulos, P. Legg, Eficient and interpretable real-time malware detection
using random-forest, in: 2019 International conference on cyber situational awareness,
data analytics and assessment (Cyber SA), IEEE, 2019, pp. 1–8.
[16] B. Marais, T. Quertier, C. Chesneau, Malware analysis with artificial intelligence and a
particular attention on results interpretability, in: Distributed Computing and Artificial
Intelligence, Volume 1: 18th International Conference 18, Springer, 2022, pp. 43–55.
[17] J. Dolejš, M. Jureček, Interpretability of machine learning-based results of malware
detection using a set of rules, in: Artificial Intelligence for Cybersecurity, Springer, 2022,
pp. 107–136.
[18] A. Orlenko, J. H. Moore, A comparison of methods for interpreting random forest models of
genetic association in the presence of non-additive interactions, BioData Mining 14 (2021) 9.</p>
      <p>URL: https://doi.org/10.1186/s13040-021-00243-0. doi:10.1186/s13040-021-00243-0.
[19] H. Xu, X. Zhang, H. Li, G. Xiang, An ensemble of adaptive surrogate models based on
local error expectations, Mathematical Problems in Engineering 2021 (2021) 8857417. URL:
https://doi.org/10.1155/2021/8857417. doi:10.1155/2021/8857417.
[20] F. Fumagalli, M. Muschalik, E. Hüllermeier, B. Hammer, Incremental permutation
feature importance (ipfi): towards online explanations on data streams, Machine Learning
112 (2023) 4863–4903. URL: http://dx.doi.org/10.1007/s10994-023-06385-y. doi:10.1007/
s10994-023-06385-y.
[21] M. T. Ribeiro, S. Singh, C. Guestrin, " why should i trust you?" explaining the predictions
of any classifier, in: Proceedings of the 22nd ACM SIGKDD international conference on
knowledge discovery and data mining, 2016, pp. 1135–1144.
[22] S. M. Lundberg, S.-I. Lee, A unified approach to interpreting model
predictions, in: I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S.
Vishwanathan, R. Garnett (Eds.), Advances in Neural Information Processing Systems</p>
    </sec>
    <sec id="sec-8">
      <title>A. Appendix</title>
      <sec id="sec-8-1">
        <title>A.1. Experimental Setting and Details</title>
        <p>Dataset and Pre-processing. The Elastic Malware Benchmark for Empowering Researchers
(EMBER), released in 2018 [9], is a well-known dataset of malware samples. The EMBER dataset
is the main dataset used throughout this study and provides a comprehensive collection of
features extracted from Windows Portable Executable (PE) files. The dataset comprises both
benign and malicious samples. It contains features from 1.1 million PE files with diverse attack
types, of which 800, 000 are labelled samples (400, 000 benign and 400, 000 malicious), and
300, 000 are unlabelled samples. We harnessed only the labelled samples for our study.</p>
        <p>
          While the EMBER dataset is in JSON format, for our experimental analysis we utilized the
version with derived features, as defined by Mojžiš and Kenyeres [
          <xref ref-type="bibr" rid="ref11">25</xref>
          ]. This dataset consists
of binary features (each feature can be either true or false – i.e. it is boolean, denoting the
presence/positiveness or the absence/negativeness of each feature) to create a simplified
representation. This representation is actually a variation of the ontology realized over the EMBER
dataset by Švec et al. [
          <xref ref-type="bibr" rid="ref12">26</xref>
          ].
        </p>
        <p>
          Feature Selections Methods. Since the majority of interpretable methods have severe
performance limitations and also fail at providing human-readable explanations when a large
feature set is available, we compared our approach against other interpretable models using the
same feature selection techniques used in the original papers. In particular, following what was
done in [
          <xref ref-type="bibr" rid="ref17 ref18">31, 32</xref>
          ], a decision tree-based feature selection technique was used in the experiments
to identify and retain the most informative features. In the experiments we considered a varying
amount of the most informative features. On the other hand, for the comparison of LENs with
state-of-the-art black-box models, like Deep Neural Networks (DNN), we used the full of set of
features.
        </p>
        <p>
          Evaluation Metrics. We evaluated both the LEN model and explanations performance using
standard metrics, i.e. accuracy, precision, recall, False Positive Rate and F1-score. For evaluating
the explanations performance, two additional metrics were employed: Fidelity [
          <xref ref-type="bibr" rid="ref19 ref4">4, 33</xref>
          ] and
Complexity [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ].
        </p>
        <p>The Fidelity metric measures the extent to which explanations faithfully represent the inner
workings of predictive models. Formally, given a data collection, a predictive model (ModelPM),
and a model explanation (ModelEx), the Fidelity(ModelEx) is defined as the accuracy obtained
when comparing the predictions made by ModelPM and the predictions derived from the
explanations ModelEx:</p>
        <p>Fidelity =</p>
        <p>1 ∑︁ Acc (ModelPM(), ModelEx())
 =1
(1)
where  is the number of samples in the data collection, Acc denotes the accuracy metric,
and ModelPM() and ModelEx() represent the predictions made by ModelPM and ModelEx,
respectively, for the -th sample . This fidelity metric will serve as a crucial indicator of the
trustworthiness of the explanations extracted.</p>
        <p>The Complexity metric counts the number of terms in the explanation as a proxy for the
human understandability of the explanation.</p>
      </sec>
      <sec id="sec-8-2">
        <title>A.2. Global Explanations in Tailored LENs</title>
        <p>Algorithm 1 reports the full procedure to get the global explanations for Tailored LENs.</p>
        <p>Algorithm 1: Compilation of the global explanations in Tailored LENs.</p>
        <p>Input: _,  _ , 
 ℎℎ ←  _ 
_ ← FilterExpl(_,  ℎℎ)
 ← EvaluateAcc(_)
while not reached optimal accuracy do
 ℎℎ ←  ℎℎ − 
_ ← FilterExpl(_,  ℎℎ)
 ← EvaluateAcc(_)
if  &gt;  then
_ ← _
 ← 
else</p>
        <p>break
 ℎℎ ←  _ 
while not reached optimal accuracy do
 ℎℎ ←  ℎℎ + 
_ ← FilterExpl(_,  ℎℎ)
 ← EvaluateAcc(_)
if  &gt;  then
_ ← _
 ← 
else</p>
        <p>break // Optimal accuracy reached</p>
      </sec>
      <sec id="sec-8-3">
        <title>A.3. Human-expert remarks on logic explanations provided by LENs.</title>
        <p>
          Table 3 shows some local explanations provided by LENs together with some remarks provided
by a cybersecurity expert. The expert highlighted that all the explanations indicated meaningful
reasons for the sample being a malware, and that it is impressive to be able to have this level
of insight into the workings of an ML-based model that was able to process the full EMBER
dataset. At the same time, all explanations were more general and abstract compared to those
derived by concept learning on a fractional dataset [
          <xref ref-type="bibr" rid="ref10">24</xref>
          ].
has_section_high_entropy ∧
¬is_dll ∧ ¬has_debug
has_write_execute_section ∧
¬has_debug
has_section_high_entropy ∧
¬has_signature
has_section_high_entropy ∧
sect_text_write_execute_section
        </p>
        <p>Cybersecurity expert remarks
It points to packed malware that is not
.dll and has not debug symbols enabled
(which is a typical malware behaviour).</p>
        <p>Malware can typically use a section with
write and execute permission for self
injection.</p>
        <p>This also points to packed malware and
malware usually has no signature
It points to packed (encrypted) code</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Ye</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Adjeroh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Iyengar</surname>
          </string-name>
          ,
          <article-title>A survey on malware detection using data mining techniques</article-title>
          ,
          <source>ACM Computing Surveys</source>
          <volume>50</volume>
          (
          <year>2017</year>
          )
          <fpage>1</fpage>
          -
          <lpage>40</lpage>
          . doi:
          <volume>10</volume>
          .1145/3073559.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>H. R.</given-names>
            <surname>Borojerdi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Abadi</surname>
          </string-name>
          ,
          <article-title>Malhunter: Automatic generation of multiple behavioral signatures for polymorphic malware detection</article-title>
          ,
          <source>in: ICCKE</source>
          <year>2013</year>
          , IEEE,
          <year>2013</year>
          , pp.
          <fpage>430</fpage>
          -
          <lpage>436</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>D.</given-names>
            <surname>Venugopal</surname>
          </string-name>
          , G. Hu,
          <article-title>Eficient signature based malware detection on mobile devices</article-title>
          ,
          <source>Mobile Information Systems</source>
          <volume>4</volume>
          (
          <year>2008</year>
          )
          <fpage>33</fpage>
          -
          <lpage>49</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>G.</given-names>
            <surname>Ciravegna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Barbiero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Giannini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gori</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Liò</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Maggini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Melacci</surname>
          </string-name>
          ,
          <article-title>Logic explained networks</article-title>
          ,
          <source>Artificial Intelligence</source>
          <volume>314</volume>
          (
          <year>2023</year>
          )
          <article-title>103822</article-title>
          . URL: https:// www.sciencedirect.com/science/article/pii/S000437022200162X. doi:https://doi.org/ 10.1016/j.artint.
          <year>2022</year>
          .
          <volume>103822</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>G.</given-names>
            <surname>Ciravegna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Giannini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Barbiero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gori</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Lio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Maggini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Melacci</surname>
          </string-name>
          ,
          <article-title>Learning logic explanations by neural networks</article-title>
          ,
          <source>in: Compendium of Neurosymbolic Artificial Intelligence</source>
          , IOS Press,
          <year>2023</year>
          , pp.
          <fpage>547</fpage>
          -
          <lpage>558</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>P.</given-names>
            <surname>Barbiero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Ciravegna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Giannini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Lió</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gori</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Melacci</surname>
          </string-name>
          ,
          <article-title>Entropy-based logic explanations of neural networks</article-title>
          ,
          <source>in: Proceedings of the AAAI Conference on Artificial Intelligence</source>
          , volume
          <volume>36</volume>
          ,
          <year>2022</year>
          , pp.
          <fpage>6046</fpage>
          -
          <lpage>6054</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>R.</given-names>
            <surname>Jain</surname>
          </string-name>
          , G. Ciravegna,
          <string-name>
            <given-names>P.</given-names>
            <surname>Barbiero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Giannini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Bufelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Lio</surname>
          </string-name>
          ,
          <article-title>Extending logic explained networks to text classification</article-title>
          ,
          <source>in: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>8838</fpage>
          -
          <lpage>8857</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S.</given-names>
            <surname>Azzolin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Longa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Barbiero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Lio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Passerini</surname>
          </string-name>
          , et al.,
          <article-title>Global explainability of gnns via logic combination of learned concepts</article-title>
          ,
          <source>in: ICLR</source>
          <year>2023</year>
          , online,
          <year>2023</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>19</lpage>
          . 30,
          <string-name>
            <surname>Curran</surname>
            <given-names>Associates</given-names>
          </string-name>
          , Inc.,
          <year>2017</year>
          , pp.
          <fpage>4765</fpage>
          -
          <lpage>4774</lpage>
          . URL: http://papers.nips.cc/paper/ 7062-a
          <article-title>-unified-approach-to-interpreting-model-predictions</article-title>
          .pdf.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>S.</given-names>
            <surname>Nembrini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I. R.</given-names>
            <surname>König</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. N.</given-names>
            <surname>Wright</surname>
          </string-name>
          ,
          <article-title>The revival of the gini importance?</article-title>
          ,
          <source>Bioinformatics</source>
          <volume>34</volume>
          (
          <year>2018</year>
          )
          <fpage>3711</fpage>
          -
          <lpage>3718</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>P.</given-names>
            <surname>Švec</surname>
          </string-name>
          , Š. Balogh,
          <string-name>
            <given-names>M.</given-names>
            <surname>Homola</surname>
          </string-name>
          , J. Kl'uka, T. Bisták,
          <article-title>Semantic data representation for explainable windows malware detection models</article-title>
          ,
          <source>arXiv preprint arXiv:2403.11669</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>J.</given-names>
            <surname>Mojžiš</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kenyeres</surname>
          </string-name>
          ,
          <article-title>Interpretable rules with a simplified data representation - a case study with the ember dataset</article-title>
          , in: R. Silhavy, P. Silhavy (Eds.),
          <source>Data Analytics in System Engineering</source>
          , Springer International Publishing, Cham,
          <year>2024</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>10</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>P.</given-names>
            <surname>Svec</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Balogh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Homola</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kluka</surname>
          </string-name>
          ,
          <article-title>Knowledge-based dataset for training PE malware detection models</article-title>
          ,
          <source>CoRR abs/2301</source>
          .00153 (
          <year>2023</year>
          ). URL: https://doi.org/10.48550/arXiv.2301. 00153. doi:
          <volume>10</volume>
          .48550/ARXIV.2301.00153.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>C.</given-names>
            <surname>Connors</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Sarkar</surname>
          </string-name>
          ,
          <article-title>Machine learning for detecting malware in pe files</article-title>
          ,
          <source>in: 2023 International Conference on Machine Learning and Applications (ICMLA)</source>
          , IEEE,
          <year>2023</year>
          , pp.
          <fpage>2194</fpage>
          -
          <lpage>2199</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>S.</given-names>
            <surname>Lad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Adamuthe</surname>
          </string-name>
          ,
          <article-title>Improved deep learning model for static pe files malware detection and classification</article-title>
          ,
          <source>International Journal of Computer Network and Information Security</source>
          <volume>14</volume>
          (
          <year>2022</year>
          )
          <fpage>14</fpage>
          -
          <lpage>26</lpage>
          . doi:
          <volume>10</volume>
          .5815/ijcnis.
          <year>2022</year>
          .
          <volume>02</volume>
          .02.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>S.</given-names>
            <surname>Pramanik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Teja</surname>
          </string-name>
          , Ember
          <article-title>- analysis of malware dataset using convolutional neural networks</article-title>
          ,
          <source>in: 2019 Third International Conference on Inventive Systems and Control (ICISC)</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>286</fpage>
          -
          <lpage>291</lpage>
          . doi:
          <volume>10</volume>
          .1109/ICISC44355.
          <year>2019</year>
          .
          <volume>9036424</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>E.</given-names>
            <surname>Raf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Fleshman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Zak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. S.</given-names>
            <surname>Anderson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Filar</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>McLean, Classifying sequences of extreme length with constant memory applied to malware detection</article-title>
          ,
          <year>2020</year>
          . arXiv:
          <year>2012</year>
          .09390.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>H.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Cocea</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Ding</surname>
          </string-name>
          ,
          <article-title>Decision tree learning based feature evaluation and selection for image classification</article-title>
          ,
          <source>in: 2017 International Conference on Machine Learning and Cybernetics (ICMLC)</source>
          , volume
          <volume>2</volume>
          ,
          <year>2017</year>
          , pp.
          <fpage>569</fpage>
          -
          <lpage>574</lpage>
          . doi:
          <volume>10</volume>
          .1109/ICMLC.
          <year>2017</year>
          .
          <volume>8108975</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>J.</given-names>
            <surname>An</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Kilmartin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Young</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Deed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <article-title>Decision trees as feature selection methods to characterize the novice panel's perception of pinot noir wines</article-title>
          ,
          <year>2023</year>
          . URL: https://doi. org/10.21203/rs.3.rs-
          <volume>2650497</volume>
          /v1. doi:
          <volume>10</volume>
          .21203/rs.3.rs-
          <volume>2650497</volume>
          /v1.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>A.</given-names>
            <surname>Papenmeier</surname>
          </string-name>
          , G. Englebienne,
          <string-name>
            <given-names>C.</given-names>
            <surname>Seifert</surname>
          </string-name>
          ,
          <article-title>How model accuracy and explanation fidelity influence user trust</article-title>
          ,
          <year>2019</year>
          . arXiv:
          <year>1907</year>
          .12652.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>