<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>V. Kovtun);</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Information Systems and Networks Department, Lviv Polytechnic National University</institution>
          ,
          <addr-line>Stepan Bandera Street, 12, Lviv, 79013</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Vasyl' Stus Donetsk National University</institution>
          ,
          <addr-line>600-richchya Str., 21, Vinnytsia, 21000</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Viacheslav Kovtun</institution>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Vinnytsia National Technical University</institution>
          ,
          <addr-line>Khmelnytske shose, 95, Vinnytsia, 21021</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0001</lpage>
      <abstract>
        <p>The article proposes an explainable model based on a δ-relaxed formal structure, which provides a transparent interpretation of classification decisions through the ontological structure of δ-concepts. The theoretical basis of the model is a δ-modified formal conceptual analysis with support for partially implemented features and fuzzy relationships between objects and attributes. An aggregated interest function is introduced, focused on optimising the semantic consistency of δ-concepts. The model is implemented as a classifier with an explanatory layer based on δ-concepts, tested on a corpus of Ukrainian utterances in the task of automatic recognition of pragmatic types. The model demonstrated high efficiency: F1-score - 0.84, average Lift - 1.23, Δ-Stability - 0.77, label entropy - 0.50. Statistical analysis showed a significant advantage of the δ-model in terms of Lift (p = 0.049) compared to CBM, which confirms more effective detection of informative concepts without loss of accuracy. The practical significance of the study lies in the creation of interpretable models for chatbots, educational systems, and legal analysis.</p>
      </abstract>
      <kwd-group>
        <kwd>δ-relaxed formal model</kwd>
        <kwd>explainable architecture</kwd>
        <kwd>concept lattice</kwd>
        <kwd>interest function</kwd>
        <kwd>pragmatic sentence classification</kwd>
        <kwd>semantic features</kwd>
        <kwd>interpretability in NLP</kwd>
        <kwd>statistical significance 1</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>The interpretability of AI decisions is a critical condition for the implementation of language
technologies in areas of increased social and legal responsibility [1-3]. In such areas as e-justice,
medical consulting, education, and moderation of public discussions, the classification decision must
be not only accurate, but also transparent for the end user – a lawyer, doctor, teacher, or moderator.
For example, in the analysis of court decisions to identify language patterns of bias, it is necessary
to explain which language constructs were the basis for the classification. In medical chatbots, it is
essential not only to provide an answer but also to argue why a particular directive is interpreted as
a request, and not as an independent decision. In educational platforms, the classification of the type
of student statement (question, statement, doubt) affects the adaptation of academic content, which
requires formal justification of the model's actions.</p>
      <p>Despite technical progress in the field of transformative architectures, most language models do
not provide the ability to trace which features (explicit, implicit, or partial) influenced the
classification [4-6]. Natural language is not reduced to complete binary relationships between words
and meanings: it is common to encounter cases of incomplete or implicit implementation of semantic
features, as well as context-dependent variations. It gives rise to an open scientific problem: how to
build an explainable architecture that combines flexibility in feature representations with formal
transparency of decisions, ensuring interpretation even in cases of unclear implementation of
linguistic information. There is a need to create an explainable model that will provide an explanation
of classification decisions in a formalised form. The focus of the study is the classification of
Ukrainian utterances by pragmatic types (statements, directives, questions, expressives), within
which the explanation of communicative intention requires not only identification but also structural
justification.</p>
      <p>The explainability of classification decisions in natural language processing is critically important
in the context of pragmatic utterance analysis. Existing classification architectures often demonstrate
high accuracy, but remain opaque about which features influenced the assignment of an utterance
to a particular pragmatic type. It is especially complicated in cases of partial, implicit, or variable
implementation of linguistic features. We will conduct a systematic critical analysis of modern
approaches to explaining the decisions of NLP models.</p>
      <p>Modern large language models (GPT-4, Claude, Gemini) implement explainability through textual
generation of reasoning, driven by special instructions (prompt-based explanation). The most
common methods are Chain-of-Thought prompting [7], Self-Generated Rationales [8], and
Instruction Fine-Tuning [9]. They are used to explain classification decisions in the form of
sequential logical phrases, in particular when defining pragmatic types of statements (questions,
statements, directives). Although such explanations are intuitive, they are uncontrolled, simulative
in nature and are not based on the deterministic logic of the model. In most cases, they depend on
patterns learned during training, rather than on interpreted internal representations. Variable results
when reformulating a query, instability of rationalisation, and lack of reproducible formalised
structure limit their application in tasks where transparent and controlled interpretation is required.</p>
      <p>Post-facto interpretation of deep learning results in NLP is most often implemented by LIME [10],
SHAP [11], Integrated Gradients [12] and Attention Rollout [13] methods. These approaches provide
a local assessment of the importance of features or tokens, based on gradient influence or
approximation of the original model by linear interpreters. Their advantage is fast integration with
non-interpretable models, such as BERT, RoBERTa, and DistilBERT, without modifying the
architecture. For example, in classifying statements as directives or questions, these methods can
determine which tokens influenced the decision. However, the explanations are unstable when
paraphrasing, do not have a common logical structure, and do not guarantee compliance with the
syntactic-semantic implementation of the features. In addition, such methods do not allow the
detection of implicitly implemented or partial features, which is critical in pragmatic analysis.</p>
      <p>Formal rule-based approaches are implemented in the form of decision trees, IF–THEN patterns,
and rule-based frameworks such as SlugNERDS [14] and RuleBERT [15]. In such models, each
classification decision is accompanied by a precise sequence of logical conditions, which allows for
full decision tracing. The advantage is high interpretability and control over the classification logic,
which will enable you to accurately determine which linguistic features, for example, the presence
of an imperative mood or an interrogative pronoun, became the basis for attributing a sentence to a
specific pragmatic type. At the same time, such models are rigid to variable language constructions
and are unable to adequately process statements with implicit or partial implementation of features.
The scalability of rule-based systems to open-domain environments or unstructured corpora is
limited.</p>
      <p>Explanation methods using graph representations are based on the construction of semantic
graphs or Graph Attention Networks (GAT) [16], which allow interpreting model solutions through
the relationships between concepts. Knowledge graph-based reasoning (e.g., KG-XAI [17]) and
pathbased explanations are also used in NLP tasks to form reasoning trajectories. The advantage is the
ability to visualise complex dependencies, integration with external ontologies (ConceptNet,
FrameNet [18]), and the use of relationships between concepts as a basis for explaining classification.
However, the application of such approaches to sentence analysis in natural language is complicated:
the construction of graphs requires a clear knowledge structure, while statements can have implicit
or fuzzy semantics. In addition, there is no direct connection between the vertices of the graph and
the grammatical structure of the sentence, which makes it impossible to fully trace solutions.</p>
      <p>The Formal Concept Analysis (FCA) method [19] is used to construct ontological structures based
on binary relations between objects and features. In the field of NLP, FCA is used for thematic
classification, generalisation of semantic constructs, and formation of concept lattices [20, 21]. The
main advantage is the ability to construct an interpreted concept lattice that describes sets of objects
with the same features. It allows classification logic to be formalised in the form of explainable rules,
as well as generalised statements by semantically similar features. However, classical FCA works
with a rigid binary matrix that assumes full implementation of features in each object. In cases of
partial, variable, or implicit implementation of features, which is typical for pragmatic sentences,
such a model loses relevance or requires excessive discretisation, which leads to the loss of
meaningful information.</p>
      <p>Therefore, in the context of analysing the advantages and disadvantages of classical approaches
to the explainable classification of pragmatic statements, the use of δ-relaxed models with built-in
interest functions is promising. Post-factum methods and generative LLM tools provide only local
simulated explanation, rule-based and FCA approaches are rigid with respect to partially realised
features, and graph-based solutions require external ontologies and do not formalise the logic of
reasoning. Against this background, δ-relaxed formalisation allows combining structural
interpretability, fuzziness, and explainability, which makes it a relevant basis for building explainable
architectures in the tasks of semantic-pragmatic analysis of natural language texts.</p>
      <p>The object of the research is the process of explainable classification of pragmatic statements in
natural language with partially implemented semantic features.</p>
      <p>The subject of the research is the theoretical foundations, formal models and methods of
constructing an interpreted classification of pragmatic statements in natural language, taking into
account the partial implementation of semantic features. It includes the analysis of modern
explainable approaches (prompt-based, post-hoc, rule-based, graph-based, FCA) and the
development of a δ-relaxed conceptual model with interest functions, which explains decisions based
on the semantic and grammatical characteristics of sentences.</p>
      <p>The research aims to improve the interpretability and flexibility of the process of classifying
pragmatic statements by developing a δ-relaxed model capable of formally reflecting the partial
implementation of semantic features in a natural language corpus based on conceptual structures
and interest functions.</p>
      <p>The article is structured as follows: Section 2 presents the theoretical basis of the study: a
formalisation of the δ-relaxed model of explanatory classification of pragmatic statements is carried
out, which takes into account the partial implementation of semantic features in the natural language
environment. A generalised incident relation is proposed, formal definitions of δ-concepts, interest
functions (in particular, target entropy, Δ-stability, Lift), as well as analytical assessments that
provide a ranking of generalisations according to their explanatory potential, are proposed. Section
3 presents the results of the experimental study: a corpus of pragmatic statements is described, the
construction of δ-relaxed lattices is implemented, interest analysis is performed, and examples of
interpretations of classification solutions with numerical quality assessments (measures of
generalisation, information gain, stability, etc.) are demonstrated. Section 4 summarises the
conclusions of the study: the scientific novelty is highlighted, the effectiveness and practical value
of the proposed model are confirmed, its limitations are identified, and promising directions for
further development are outlined.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Models and Methods</title>
      <p>Classical FCA is a powerful tool for detecting hidden categories in tabular structures. However, its
application in computational linguistics is limited by a rigid binary logic: either a feature is present
or it is not. In contrast, in natural language, features can be partially realised, with vagueness or
instability of detection, for example, in the case of co-occurrence structures (statistically significant
co-occurrences of words, such as "sharp criticism"), synonymous variations, or pragmatic shifts.
Therefore, there is a need for a δ-relaxed model, where the correspondence between an utterance
and a feature allows for a gradual degree of membership, controlled by the parameter δ. In corpus
linguistics, many features do not have the rigid binary nature predicted by classical FCA. Categories
such as modality, pragmatic shift, co-occurrence relevance, or synonymy are not detected with
absolute certainty, but with a certain degree of probability or fuzziness. It necessitates a δ-relaxed
model that allows the use of the degree of membership of a feature to an utterance within the interval
[0,1], which allows for accurate modelling of fuzzy or partial manifestations of linguistic properties.</p>
      <p>
        In corpus linguistics, a set of objects  is considered a collection of tokenised sentences or
contexts, utterances or fragments of discourse. The set of features  includes morphological
categories, part-of-speech tags, syntactic relations, semantic labels (e.g. modality, movement,
agentivity) or co-occurrence patterns. Formal context is defined as
 = ( ,  ,  ),  ⊆  ×  ,
(
        <xref ref-type="bibr" rid="ref1">1</xref>
        )
where ( ,  ) ∈  if and only if the feature  is inherent in the statement  with high confidence.
For fuzzy data, a δ-relaxed version (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) is introduced:  = ( ,  ,  ),  ⊆  ×  × [0,1], where
 ( ,  ) ∈ [0,1] is the degree of correspondence of the feature  to the sentence  . In general, the
context  or  (in the case of the δ-model) represents the relationship between a set of sentences
(statements) and a set of features (lexical, morphological, semantic, syntactic, etc.). This structure
corresponds to a sparse table or matrix, where the values can be both binary (0/1) and gradational
(fuzzy), which will be reflected in the examples in Section 3.
      </p>
      <p>Closure operators allow us to associate sets of sentences with corresponding sets of features:  ′ =
{ ∈  |∀ ∈  : ( ,  ) ∈  },  ⊆  ,  ′ = { ∈  |∀ ∈  : ( ,  ) ∈  },  ⊆  . In the δ-context, fuzzy
closure operators are introduced:  = { ∈  | ( ,  ) ≥  , ∀ ∈  },  ′ = { ∈  | ( ,  ) ≥
 , ∀ ∈  }.</p>
      <p>
        On the basis of classical FCA (i.e. over  ), the formal concept is interpreted as a pair ( ,  ), which
satisfies the two-way closure condition:
 ′ =  ,  ′ =  ,  ⊆  ,  ⊆  .
(
        <xref ref-type="bibr" rid="ref2">2</xref>
        )
      </p>
      <p>
        It means that all sentences with  have standard features of  , and vice versa –  characterises
only these sentences. In the following, unless otherwise stated, all pairs ( ,  ) are interpreted relative
to a fixed context  or  . In δ-cases, condition (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) may have a fuzzy implementation.
      </p>
      <p>The generalisation relation can order the concepts:</p>
      <p>
        ( ,  ) ≺= ( ,  ) ⇔  ⊆  ,  ⊆  . (
        <xref ref-type="bibr" rid="ref3">3</xref>
        )
      </p>
      <p>Thus, concepts with smaller scope and more specialised features are concretisations of more
general ones. As the corpus size or number of features increases, a combinatorial explosion of
concepts is observed. Therefore, there is a need for cognitively oriented filtering methods based on
interest indices, which we will consider later.</p>
      <p>The coverage ratio in a lattice is given by:</p>
      <p>
        ( ,  ) ≺ ( ,  ) ⇔  ⊂  ,  ⊂  , ∄( ,  ):  ⊂  ⊂  . (
        <xref ref-type="bibr" rid="ref4">4</xref>
        )
In the neural network interpretation, expression (
        <xref ref-type="bibr" rid="ref4">4</xref>
        ) corresponds to a direct connection between
nodes (neurons) that implement the transition from a general pattern to a more specific one. In
practical cases, concepts that have only a partial overlap of the volumes  ∩  ≠ ∅, but are not in
a covering relation, are often encountered. To ensure the coherence of the δ-lattice, δ-bridges are
introduced - auxiliary concepts that connect structurally close, but formally uncovered pairs:
 , ,  , : = 
∩  , (
∩  )′ ,
(
        <xref ref-type="bibr" rid="ref5">5</xref>
        )
where  , – is formed by the δ-closure of the intersection  , . However, the reverse closure may
not be fulfilled. Such pairs may not satisfy the strict conditions of a formal concept, but they play the
role of connecting units in δ-structures. δ-bridges of the form (
        <xref ref-type="bibr" rid="ref5">5</xref>
        ) play the role of cognitively relevant
transitional concepts that connect concepts with a partial but significant intersection of volumes.
Although such pairs do not satisfy the conditions of a formal concept in the classical sense, they
ensure the coherence of the δ-lattice and the continuity of the semantic structure between concepts.
In the context of explainable architectures, δ-bridges play the role of buffer nodes that contribute to
the understandable interpretation of the internal layers of the model. Their effectiveness will be
analysed empirically in the next section.
      </p>
      <p>In the tasks of building neural networks based on lattices of formal concepts, filtering concepts is
of particular importance, which will subsequently form the structure of hidden layers. Due to the
exponential growth of the number of concepts with an increase in the size of the input space, there
is a need to select the most relevant ones. This task is solved through interest indices - numerical
functions that rank formal concepts by their cognitive, statistical or logical significance. Let us
present the interpretation of interest indices, the most relevant for the δ-model of formal analysis of
linguistic concepts, and the most common indices.</p>
      <p>
        Basic Level Index has a cognitive motivation, a concept must have high internal coherence, be
more coherent than its superconcepts and not much less coherent than its subconcepts:
 ( ,  ) =   ( ,  ),  ( ,  ),  ( ,  ) , (
        <xref ref-type="bibr" rid="ref6">6</xref>
        )
where  (⋅) is the aggregating t-norm (for example, the product),  ( ,  ) is the coherence of the
concept itself,  ( ,  ) is the comparison with superconcepts,  ( ,  ) is the comparison with
subconcepts. Coherence is calculated as:
      </p>
      <p>∅( ,  ) = | |(| | ) ∑{ , }⊂ , sim( ,  ),
where  = { ∈  |( ,  ) ∈  } is the set of features belonging to object  , and sim( ,  ) is the
similarity between two objects is based on their features. The metrics used are:
(7)
(8)
(9)
(10)
(| ∩  | + | − ( ∪  )|)
sim ( ,  ) = ,</p>
      <p>| |
sim ( ,  ) = || ∩∪ ||,
where  ⊆  is a subset of features by which similarity is calculated.</p>
      <p>
        In the context of (7), the functions  { , , }( ,  ) mentioned in (
        <xref ref-type="bibr" rid="ref6">6</xref>
        ) are defined as:
 ( ,  ) =  ∅( ,  ),  ( ,  ) = 1 − | ( , )| ∑ ∈ ( , ) ∅∅(( , )),
      </p>
      <p>
        ( ,  ) = 1 − | ( , )| ∑ ∈ ( , ) ∅∅(( , )),
where  ( ,  ),  ( ,  ) are the sets of immediate super- and subsumed pairs ( ,  ) in the
complete lattice of formal concepts constructed over the context  (see (
        <xref ref-type="bibr" rid="ref3">3</xref>
        )). We will denote the
realisations of index (
        <xref ref-type="bibr" rid="ref6">6</xref>
        ) as  (based on SMC) and  (based on Jacquard).
      </p>
      <p>The Target Entropy index is used in classification problems where each object has a target label.
It is defined as the variance or entropy of classes among objects of a concept:
 ( ,  ) =  ( ( ),  ∈  ),
(11)
where  (⋅) is the variance, and  ( ) ∈  is the target label of an object  belonging to a set of
classes  (e.g., tonality, topic, pragmatic type). The lower the entropy, the more consistent the
concept is with a particular class.</p>
      <p>( ,  ) = 
| | −</p>
      <p>,
 ( ,  ) = ∏</p>
      <p>( )
( ) ,</p>
      <p>The classical stability of a concept characterises the stability of its features to variations of a set
of objects. The computational complexity of the process of assessing stability in practice uses the
approximate Δ-Stability index:</p>
      <p>where  = ( ,  ),  is any subconcept of the pair  such that  ∈  . The index Δ shows how
"distant" the concept is from its closest subconcepts and, accordingly, how structurally stable it is.</p>
      <p>The Lift index is used in association analysis to identify non-obvious but statistically informative
associations between features.:
(12)
(13)
where  ( ) = | | is the probability of occurrence of feature  ∈  ,  ( ) = | | is the probability
of the joint occurrence of all features  . If  ( ,  ) &gt; 1, the features from the set L are positively
correlated. To avoid numerical instability, the logarithmic version (13) is often used:
  ( ,  ) = ∑ ∈  ( ) −  ( ). (14)</p>
      <p>Among the above indices, the most potentially suitable for explainable architectures are 
(reflecting internal coherence) and Δ-Stability (providing structural separation). Their combination
allows for the formation of hidden layers of the neural network as interpreted concepts, consistent
in features, and isolated in a lattice topology.</p>
      <p>
        In explainable AI tasks for natural language processing, a critical stage is the detection of
interpreted concepts that form the structure of the hidden layers of the neural network. Unlike
tabular data, language corpora are characterised by complex semantics, uneven feature structure and
high variability, which makes it impossible to use classical association rules or rigid ontologies.
Therefore, the goal is to construct a set of δ-concepts  {( ,  )}с ⊆  , which, taking into
account the relaxation of the membership relation, cover  and correspond to high values of the
interestingness indices (expressions (
        <xref ref-type="bibr" rid="ref6">6</xref>
        )–( f14)). Here  is a natural number that denotes the number
of selected δ-concepts (concepts) in the set  ,  is a δ-lattice of formal concepts, that is, the set of
all pairs ( ,  ) that satisfy the condition of partial correspondence between objects and features with
an accuracy of at least δ:  = ( ,  )  ⊆  ,  ∈  , ∀ ∈  : |{ ∈ ||( |, )∈ }| ≥  . Each concept in the
structure of the set of δ-concepts is a potential node in the neural network model.
      </p>
      <p>In the sought-after δ-context  = ( ,  ,  ), the membership relation  ⊆  ×  is defined as a
partial correspondence: an object  ∈  satisfies the feature set  ⊆  if:
|{ ∈ |( , )∈ }| ≥  . ( (15)</p>
      <p>| |</p>
      <p>Inequality (15) must hold for each  ∈  , i.e., the features  must be characteristic of all objects
of the set  with an accuracy of at least δ.</p>
      <p>The construction of the lattice is preceded by a procedure for preprocessing the corpus, which
includes:
– tokenisation, lemmatisation, PoS-markup;
– generation of features  , including grammatical (gender, number, tense), syntactic (subject,
predicate, object) and semantic (action, place, evaluation) markers;
– construction of  taking into account the given level of fuzziness δ.</p>
      <p>Based on  , a δ-lattice of concepts  is constructed. In this case, the closures  ≈  ,  ≈  are
understood as partial, taking into account the fuzzy relation  . It allows maintaining consistency
with the basic theory of formal concepts when transitioning to a fuzzy model.</p>
      <p>
        Each concept ( ,  ) ∈  is a candidate for the role of a node of the explainable architecture. The
choice of a specific index or its combination depends on the task:
– Basic Level Index, in particular its implementations 
(based on Simpson's metric) and 
(based on Jaccard's metric). It is recommended for detecting concepts with high internal coherence
that are well-matched by features. The index is defined by formula (
        <xref ref-type="bibr" rid="ref6">6</xref>
        ), coherence components (7)–
(10), including the corresponding similarity metrics (8), (9), and aggregation is performed via t-norm.
      </p>
      <p>– Target Entropy is used when objects have target labels (classes), for example, in classification
or clustering tasks. For numerical labels, the variance is used – formula (11), for categorical labels –
Shannon's entropy [22, 23]. Concepts with low entropy are considered well-matched with a
particular class.</p>
      <p>– Δ-Stability index is effective for structural filtering of concepts. It assesses how isolated an idea
is in the lattice structure – that is, whether it "overlaps" with a large number of subconcepts. Formally
defined by formula (12). It allows us to select the most structurally stable nodes that retain their
significance when the corpus varies.</p>
      <p>– Lift reflects the statistical non-obviousness or associative strength of features: concepts with a
high Lift value reveal unexpected but significant combinations of features that occur together more
often than expected. The standard definition is formula (13), and the logarithmic form is (14). It is
helpful in detecting correlated feature patterns, particularly in corpora with a latent structure.</p>
      <p>In cases where it is necessary to balance coherence, stability and associativity, it is advisable to
use an aggregate utility function:

( ,  ) =   ( ,  ) +   ( ,  ) +  
 ( ,  ) + 
1 −  ( ,  ) ,
which combines the corresponding indices with weights 
∈ [0,1],  = 1,4, ∑

Suppose that we need to find a set of concepts 
= {( ,  )}
∈  that covers the corpus  =
⋃
 and maximises the total utility:

{( , )}
∑</p>
      <p>( ,  ).</p>
      <p>Overlap between sets</p>
      <p>is allowed, since objects can be relevant to several concepts - this
increases the accuracy and flexibility of the explainable model. Each concept ( ,  ) must
correspond to a node of the hidden layer, which is activated on the set  , responds to features 
and has an interpreted linguistic representation. Nodes are connected according to the partial order
of the δ-lattice: if ( ,  ) ≤  ,</p>
      <p>, then a directed connection is formed between the corresponding
neurons. Both direct and transitive connections are allowed, creating a multi-level generalisation
hierarchy. The δ-lattice model formalised in this way forms the basis for an explainable neural
network, which relies on linguistically interpreted nodes and a topologically ordered structure.</p>
      <p>We will conclude the section by formulating the concept of constructing an explainable neural
network based on the δ-lattice of formal concepts І , formed by filtering the full lattice ( ,  ,  )
according to the interest function</p>
      <p>( ,  ), defined according to expression (16) as a weighted
linear combination of the indices  ,  ,</p>
      <p>and 1 −  . The coefficients  ,  = 1,4, which specify
the weight of each index, are selected according to the type of problem. In particular, for classification
models, priority was given to Target Entropy, while for semantic analysis, the coherence index  ,
and for knowledge generalisation, Lift.</p>
      <p>Each concept ( ,  ) ∈</p>
      <p>was considered as a formal cognitive unit that associates a subset of
objects</p>
      <p>
        ⊆  with a set of features  ⊆  , where features are grammatical, semantic, or pragmatic
characteristics relevant to the target classification task. The distribution of features was modeled as
a phase-Markov process with absorption (see expression (
        <xref ref-type="bibr" rid="ref5">5</xref>
        )), which allowed us to reflect the duration
of the appearance of features, delay, disappearance, as well as the probabilistic sequence of their
activation in time, which is critically important for the analysis of speech or discursive texts [24].
      </p>
      <p>The selection of a subset of concepts that directly shape the architecture of the model is
interpreted as a corpus coverage problem, generalised by expression (17). It guarantees complete
coverage of the set  by concepts with 
= {( ,  )}
⊆  , while simultaneously maximising the
(16)
= 1.
(17)
total significance according to  ( ,  ). The overlap between the sets  is not limited, which
increases the flexibility of the interpreted model.</p>
      <p>Each node of the hidden layer corresponds to a specific concept ( ,  ) ∈  and is activated if the
input object  ∈  belongs to the set  , and at least one feature from  is present in the context. To
form vector representations of concepts, embeddings of features from  are used, obtained using a
pre-trained transformer model such as BERT or RoBERTa. Each concept ( ,  ) is identified with the
averaged vector of its features. On this basis, keys and attention values of the scaled dot-product type
are formed:</p>
      <p>Attention( ,  ,  ) = softmax
 ,
(18)
where  is the query vector formed on the basis of the input object  ,  is the key matrix
containing the vector representations of the concepts ( ,  );  is the value matrix associated with
the concepts or their context,  is the number of components in the key or query vector, i.e. the
dimension of the space in which the comparison between  and  is made. Attention weights not
only optimise predictions, but also act as a means of explanation: each coefficient in the attention
matrix is interpreted as a measure of semantic or functional proximity between concepts that were
activated simultaneously. This approach allows us to reconstruct the logic of the classification
decision at the level of the explained structures.</p>
      <p>In cases where the sets  and  of two concepts partially overlap, aggregated (synthetic)
concepts of the form  ∩  ,  ∪  should be introduced into the model, which provides
consistent coverage of objects without losing relevant features, while maintaining logical
consistency between nodes.</p>
      <p>In the semi-supervised learning mode (when part of the corpus objects is not manually annotated),
it is advisable to use the loss function:
on learning (at  = 0 we have pure supervised learning); 
=
 = ∑ ∈  ( ,  ̂ ) +  ∑ ∈    ̃ ,  ̂ , (19)
where  ⊆  is the set of labelled objects: objects  for which there is an accurate label  ;  ⊆ 
is the set of unlabeled objects: objects  for which the actual label is unknown, but there is a
predicted/artificial label  ̃ ;  is the actual target label of the object  ∈  , which can be categorical
or numeric (e.g., class, topic, tone);  ̂ ,  ̂ are the model predictions for objects from sets  and  ,
respectively (the output values of the neural network);  ( ,  ) is the cross-entropy loss function
between label  and prediction  ; it measures how much the prediction  ̂ deviates from the actual or
pseudo label;  ≥ 0 – hyperparameter that specifies the degree of influence of unlabeled examples
,
– confidence in concept 
according to the  index (12). The set  formed according to criterion (17) provides complete
coverage of the corpus, and also forms an interpreted multilayer neural network architecture, where
each node corresponds to a logically justified formal concept, and attention connections allow us to
trace the semantic logic of the model.</p>
      <p>In general, δ-relaxed contexts and corresponding formal concepts can be used as explainable
nodes in modern neural network architectures. The proposed lattice structure is logically consistent
with approaches such as Concept Bottleneck Models, where each concept represents a separate
cognitive feature that can be interpreted independently. δ-bridges, in this context, form semantically
justified connections between such nodes, allowing the modelling of the transition between different
levels of generalisation. The architecture formed in this way will be characterised by the openness
of the internal logic of decision-making, which is key for explainable AI in the field of natural
language processing.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Results and Discussion</title>
      <p>As a result of the research presented in Section 2, a hypothesis was formulated according to which
the explainable model, built on the author's δ-relaxed formal structure, provides a higher level of
interpretability while maintaining classification efficiency at the level of modern models, such as
Classical FCA, Concept Bottleneck Models (CBM), ProtoPNet, and SHAP/LIME. In this context,
interpretability is understood as a set of features that includes the presence of a formalised
ontological structure of concepts, their stability to corpus variations, semantic consistency of features
within an idea, mapping of the target label with low entropy, and confirmation of conceptual
relevance through expert cognitive validation. The structure and content of this section are focused
on experimental verification of this hypothesis.</p>
      <p>To empirically test the effectiveness of the δ-relaxed approach in explainable natural language
modelling tasks, the open corpus UD_Ukrainian-IU (Universal Dependencies version 2.13) was used,
which contains 7060 sentences with a total volume of about 122 thousand tokens. The corpus
represents the Ukrainian language in its real functioning - covering literary, journalistic, legal and
socially oriented texts. Such genre heterogeneity creates a favourable environment for identifying
the structural stability of δ-concepts to semantic and pragmatic variations. Within the framework of
the experiment, the corpus was used to implement the task of classifying sentences by the pragmatic
type of utterance. Such types include statements, questions, directives, expressives and other speech
acts that have practical significance in applied NLP scenarios.</p>
      <p>To ensure reproducibility of the results, the corpus was divided into three parts: 70% was used for
training, 15% for validation of the δ parameter and concept selection, and another 15% for final model
testing. The division was done with a fixed random seed value, which guarantees structural stability
during repeated runs and statistical comparisons.</p>
      <p>Preprocessing was implemented based on a standard UD pipeline with adaptation to the needs of
explainable modelling. Tokenisation was performed taking into account punctuation boundaries and
multi-component constructions. Lemmatisation was performed at the level of word forms with
fixation of the morpheme core. Part-of-speech PoS markup was reduced to a universal tagset with a
full morphological specification: gender, number, tense, type, degree, etc. The syntactic layer was
obtained through dependency analysis, which allowed for the automatic identification of the roles
of subject, predicate, direct and indirect object, definition, and circumstance. Semantic features were
generated semi-automatically based on heuristics and a pre-assembled lexical corpus for the
Ukrainian language. Additionally, templates were used to detect modality, evaluativeness, pragmatic
function, as well as features of action, addressee, movement, and purpose. Such characteristics, which
do not always have a straightforward syntactic implementation, are revealed indirectly – through
connotative patterns, grammatical sequences, or contextual accents – and they are the ones that are
decisive for the construction of concepts with internal cognitive coherence.</p>
      <p>For each sentence, an individual feature profile was obtained with variability within 12–22 units,
which was determined by both syntactic complexity and the degree of semantic saturation of the
statement. The average value was 17.3 features per sentence. For further processing, the formal
δcontext  = ( ,  ,  ) described in Section 2 was transformed into a numerical representation that
provides effective integration into model structures. In particular, the set of sentences  was
displayed as a feature matrix  ∈ ℝ × , where  = | |,  = | |; the set of features  formed the
columns of  , and the δ-relaxed relation  was interpreted as a weight matrix  ∈ ℝ × , which
reflected the degree of belonging of features to sentences. In addition, each object  ∈  was assigned
a target label  ∈ {0,1} in the label vector  ∈ ℝ , synchronised with the rows of  . The prepared
data were stored in the .npz format, which included the specified components: matrix  , matrix  ,
and label vector  . This structure allowed us to directly calculate interest indices, form a δ-lattice of
concepts, and implement an attention architecture.</p>
      <p>Based on the previously selected semantic-syntactic features (see section 2), the number of which
in sentences varied from 12 to 22, a δ-context was constructed in the form of a triple ( ,  ,  ). The
ratio  (,  ) = 1 was fixed if the proportion of sentences with the feature  accompanying the
glike feature vector (in the sense of the δ-distance) exceeded the given threshold δ. The parameter δ
varied within [0.5; 1.0] with a step of 0.05, and the optimality criterion was the maximum average
Lift value among concepts with Δ-stability over 0.6 under the condition of low entropy of the target
distribution. The optimal value δ = 0.85 provided a balance between generalisation, conceptual
stability and purity of segments. With this parameter, a δ-lattice of concepts was formed based on
δclosures – sets of features that stably coexist in subsets of the corpus. Connections between such
concepts are built not only under full incidence, but also under partial δ-compatibility, which
preserves a coherent topology even under conditions of fragmentary or noisy annotations. Figure 1
shows a fragment of the constructed δ-lattice of concepts, which illustrates the gradual cognitive
generalisation of features in a stable semantic subspace of the corpus. The empty concept ∅ is
interpreted as a neutral concept without semantic load.</p>
      <p>The grid in Figure 1 clearly traces the transition from basic concepts (Evaluation or Modality) to
complex cognitive configurations that include Addressee, Motion, and other relevant categories. For
example, the idea of evaluation and modality appears as a generalisation of two atomic concepts
while maintaining structural coherence within the subspace. Such a construction not only formalises
stable linguistic dependencies but also forms the basis for explicable attention mechanisms in the
tasks of classification, interpretation, and construction of trusted architectures, which, in turn, allows
us to derive formalised rules and cognitively understandable explanations.</p>
      <p>
        The construction of an explainable architecture based on δ-concepts involves the implementation
of an attention mechanism in which the key decisions of the model are based on the formal structure
of the δ-lattice of concepts. The model uses input representations obtained from the RoBERTa-base
(768 dimensions), which serve as the basis for the phase overlay of the δ-lattice of concepts  ,
generated on the basis of the δ-relaxed formal context  = (, ,  ). Each δ-concept (,  ) ∈  is
interpreted as a logical unit describing a cluster of objects  ⊆  through a feature set  ⊆  . The
pair (,  ) is δ-closed if  ( ) =  and  ( ) =  , according to definition (
        <xref ref-type="bibr" rid="ref6">6</xref>
        ). Table 1 contains
additional information supporting this process.
      </p>
      <sec id="sec-3-1">
        <title>Aggregated evaluation based on interest</title>
      </sec>
      <sec id="sec-3-2">
        <title>Quantifies the contribution of</title>
      </sec>
      <sec id="sec-3-3">
        <title>Description</title>
      </sec>
      <sec id="sec-3-4">
        <title>A δ-relaxed formal context comprising</title>
        <p>the set of objects  , the set of features  ,
and a tolerance-based incidence relation</p>
        <p>⊆  ×</p>
      </sec>
      <sec id="sec-3-5">
        <title>The set of δ-closed formal concepts</title>
        <p>( ,  ), where</p>
        <p>⊆  ,  ⊆  , and the
closure conditions  ( ) =  ,  ( ) = 
hold</p>
      </sec>
      <sec id="sec-3-6">
        <title>Formal measures reflecting stability, Lift, and target entropy of each concept</title>
      </sec>
      <sec id="sec-3-7">
        <title>Proportions of object and feature domains covered by the concept and generalisation indices</title>
      </sec>
      <sec id="sec-3-8">
        <title>Role in the Model</title>
      </sec>
      <sec id="sec-3-9">
        <title>Provides the structured input space for concept induction</title>
      </sec>
      <sec id="sec-3-10">
        <title>Encodes latent hierarchical</title>
        <p>relations
among
grouped
objects and features</p>
      </sec>
      <sec id="sec-3-11">
        <title>Guide the identification of</title>
        <p>structurally and semantically
informative concepts.</p>
        <p>Indicate the abstraction level
and potential generalisability
of concepts
δ-concepts to
explanation
the
model</p>
        <p>Component
δ-Context</p>
        <p>=
( ,  ,  )</p>
      </sec>
      <sec id="sec-3-12">
        <title>Concept Lattice</title>
      </sec>
      <sec id="sec-3-13">
        <title>Interest Indices  ,</title>
        <p>,</p>
      </sec>
      <sec id="sec-3-14">
        <title>Generalisation Indices  ,</title>
      </sec>
      <sec id="sec-3-15">
        <title>Concept Scoring Function</title>
        <p>( ,  )
interest function 
softmax 
this process.</p>
      </sec>
      <sec id="sec-3-16">
        <title>Formal Concept</title>
      </sec>
      <sec id="sec-3-17">
        <title>Scoring Function</title>
      </sec>
      <sec id="sec-3-18">
        <title>Concept Structure and Weighting</title>
      </sec>
      <sec id="sec-3-19">
        <title>Selection Threshold</title>
      </sec>
      <sec id="sec-3-20">
        <title>Applied to</title>
      </sec>
      <sec id="sec-3-21">
        <title>Filters out δ-concepts with insufficient explanatory</title>
        <p>For each δ-concept, the attention weight  ( ,  ) was calculated as the softmax of the aggregated
( ,  ), which combines the indices ⟨ ,  ,  ,  ⟩. The definition of  ( ,  ) =
( ,  ) is given in expression (16). Since the explicit form of the function 
is not
fixed; it is formed as a parameterised combination of the specified indices, with adjustable weight
coefficients that are optimised during training. Table 2 contains additional information to support</p>
      </sec>
      <sec id="sec-3-22">
        <title>Formal Concept Representation and Weighting Criteria</title>
      </sec>
      <sec id="sec-3-23">
        <title>Stability</title>
      </sec>
      <sec id="sec-3-24">
        <title>Lift</title>
      </sec>
      <sec id="sec-3-25">
        <title>Target Entropy Object Generalisation Feature</title>
        <p>Generalisation</p>
      </sec>
      <sec id="sec-3-26">
        <title>Parameter / Notation</title>
        <p>( ,  ) ∈ 
( ,  )
( ,  )</p>
      </sec>
      <sec id="sec-3-27">
        <title>Definition / Analytical Role</title>
      </sec>
      <sec id="sec-3-28">
        <title>A δ-closed pair consisting of an object subset  and a feature subset</title>
      </sec>
      <sec id="sec-3-29">
        <title>Aggregated concept relevance measure, formally defined by expression (16) contribution</title>
      </sec>
      <sec id="sec-3-30">
        <title>Explanation Metrics</title>
      </sec>
      <sec id="sec-3-31">
        <title>Proportion of subsampled contexts where the concept remains δ-closed of features within</title>
      </sec>
      <sec id="sec-3-32">
        <title>Ratio between observed and expected co-occurrence</title>
      </sec>
      <sec id="sec-3-33">
        <title>Entropy of class labels within the object set</title>
      </sec>
      <sec id="sec-3-34">
        <title>Coverage of the object domain  by the concept extent</title>
        <p />
      </sec>
      <sec id="sec-3-35">
        <title>Coverage of feature domain  by the concept intent</title>
        <p>To improve accuracy and explainability, δ-concepts were ranked by the value of  , after
which an explainable subset  ⊆  consisting of the 128 most relevant concepts was formed, the
elements of which satisfied the threshold condition of Δ-stability  ( ,  ) &gt; 0.6. The values of 
in this subset, they were normalised via softmax to form an attention profile.</p>
        <p>The final values of the  ( ,  ) attention weights were determined by the intensity of the
influence of the feature set  on the formation of the explanatory context of the sentence. In the
forward-skip phase, a weighted feature vector was formed as a convex combination of hidden
representations corresponding to δ-concepts from the subset  , with weights  ( ,  ). Thus, the
model shifted the emphasis in the semantic representation to groups of features that best explain the
decision.</p>
        <p>To reduce the impact of unstable or entropy-weak concepts, a modified loss function was
implemented, formally presented in expression (7), which took into account the δ-weighted
crossentropy deviation with weights proportional to  ( ,  ). In addition, an additional penalty was
introduced for concepts with high  or low  , which allowed automatically reducing the weight of
uninformative factors. The model was trained using the Adam optimiser, which allowed for stable
adjustment of the weight coefficients in the attention module.</p>
        <p>In the process of developing explainable architectures, it is vital to consider not only the accuracy
of prediction but also the quality of explanations: their transparency, cognitive accessibility, level of
detail, and correspondence to human ideas about semantic relationships. To critically evaluate the
proposed δ-explainable approach, it was compared with established explainable methods, in
particular SHAP, LIME, and attention-based explanations, which are actively used in transformative
architectures.</p>
        <p>The SHAP method, based on the theory of Shapley values, demonstrates a high level of
transparency in both global and local explanations. However, its computational complexity is
significant, especially when working with large models. The LIME method, which operates on local
surrogate models, is characterised by relative simplicity of implementation, but demonstrates
instability of results and weak semantic consistency. In transformer models, attention-based
explanations are common, built on the analysis of attention weights  ( ,  ), which, as shown in
formula (16) of Section 2, illustrates the distribution of attention of the model. However, such weights
do not reflect the causal relationship between features and the forecast and do not guarantee
interpretation in the strict sense, since they are intermediate internal parameters.</p>
        <p>
          In contrast to these approaches, the proposed δ-explainable architecture provides global
explainability based on formal concepts with a conceptual structure. It is based on a δ-relaxed formal
context ( ,  ,  ), where sets of sentences  and features  are connected through a δ-relaxed relation
 , which determines the partial correspondence between them. The weight matrix  , which
describes these connections quantitatively, is the basis for constructing a δ-lattice of formal concepts.
This hierarchical structure reflects semantic relations between groups of features and objects. Each
idea is formed on the basis of sets  ( ) and  ( ), according to expression (
          <xref ref-type="bibr" rid="ref1">1</xref>
          ), and generalises the
interpretation through conceptual categorisation. A qualitative comparison of the author's approach
with the closest analogues is presented in Table 3.
        </p>
        <p>In the context of statistical verification of the author's approach, the key indices of concept
interest were analysed: Lift, Δ-Stability, and Target Entropy (see expressions 11–14). These metrics
cover the main characteristics of the quality of formalised concepts in the δ-lattice: the degree of
connection between objects and features, resistance to selective influence, and the level of semantic
certainty. In particular, Lift reflects the strength of the association between sets of objects and
features; Δ-Stability is the stability of the concept to data variability; Target Entropy is the level of
uncertainty in its interpretation (the lower the entropy, the more accurate the explanation).</p>
        <p>To assess the reliability of the difference between the δ-model and the best of the considered
analogues – CBM (Concept Bottleneck Model), both the parametric t-test and the non-parametric
Mann–Whitney U-criterion were used, which provides increased reliability of the analysis in case of
deviations from normality. The assumption of normal distribution was tested using the Shapiro–
Wilk test (p &gt; 0.05 for all three metrics), which justifies the use of the t-criterion. The analysis covered
50 δ-concepts, selected according to the function  ( ,  ), which summarises the pragmatic
appropriateness of each concept in the interpreted architecture (see formula 10). Correction for
multiple comparisons was not applied due to the limited number of hypotheses and the explanatory
nature of the study. The results of the statistical analysis are presented in Table 4.</p>
        <p>Table 4 shows that the δ-model demonstrates a statistically significant improvement in Lift
compared to CBM (p = 0.049). The corresponding Cohen's d effect size for Lift is 0.51, indicating a
moderate strength of difference between the models. For Δ-Stability and Target Entropy, statistical
analysis did not confirm the significance of the difference. Still, in all cases, there was a consistent
direction of change in favour of the δ-model. It indicates a general trend towards improved quality
of formalised concepts in the δ-approach, even in the absence of strict statistical reliability.</p>
        <p>After constructing an explainable architecture based on δ-concepts, the task arose to empirically
verify the classification efficiency, the stability of the interestingness of the concepts, and the level
of interpretability of the results. First, the distribution of attention was assessed in the model, which
is based on the attention mechanism with δ-concepts as carriers of explained features. Figure 2 shows
a heat map of the level of attention activation to δ-concepts in an example from the test sample. The
horizontal axis shows the conventional designations of tokens ( –  ) corresponding to the words
of the sentence, and the vertical axis shows the indices of δ-concepts corresponding to the relevant
formal generalisations. The colour gradation scale reflects the intensity of the influence.</p>
        <p>The results presented in Figure 2 show that the model's attention is focused mainly on
semantically rich tokens with a high information load (in particular, on predicates or explicit,
emotionally coloured constructions). The presence of condensation in 5–7 concepts is explained by
their generalisation and correspondence to several contextual features at the same time. High
segmental contrast indicates the functional orientation of the attention module to meaningful
features, which is consistent with expert assessments.</p>
        <p>The next step was to study the distribution of the key measure of interest Lift, which characterises
the ratio of the actual and expected frequency of occurrence of a concept in the target class. Figure
3 shows a histogram of the distribution of Lift values among 30 selected δ-concepts. The abscissa
axis shows the Lift value in the range [0.85; 1.6], and the ordinate axis shows the frequency of
δconcepts that demonstrate the corresponding value. The red vertical line corresponds to the neutral
threshold value Lift = 1. For visualisation, binarisation was performed based on 20 intervals - some
of the concepts fall into the same binary segments.</p>
        <p>As can be seen from Figure 3, the majority of concepts exhibit Lift values above unity, indicating
their classification relevance. The presence of several concepts within 1.3–1.5 means the existence
of dominant regularities, while 2–3 concepts with Lift &lt; 1 play the role of compensators or limiters,
increasing the overall generalisation ability of the model.</p>
        <p>The analytical part of the study includes the eight most relevant δ-concepts with the parameter
Δ-Stability &gt; 0.6. Their classification metrics (accuracy, completeness, F1-value and support) are
given in Table 5. Additionally, a macro average was calculated - the average value for all concepts.
This approach was chosen due to the need to uniformly assess the effectiveness of each δ-concept
regardless of their frequency in the corpus. The metrics presented in Table 5 demonstrate the overall
balance of the model: Precision and Recall values within 0.72–0.94 not only confirm the classification
efficiency, but also indicate the absence of a significant predominance of one of the indicators, which
is a sign of harmonious learning.</p>
        <p>An in-depth analysis of the interest indices (Lift, Δ-Stability, Target Entropy), which serve as
indicators of the cognitive ability of δ-concepts, is given in Table 6. Separate values were calculated
for each δ-concept, and a generalised row was also compiled.</p>
        <p>The high values of Lift and Δ-Stability (&gt; 1.2 and &gt; 0.75, respectively) for most concepts presented
in Table 6 indicate not only their prevalence but also structural stability. At the same time, the low
entropy (0.37–0.60) confirms that the concepts are not chaotic, but are focused on a narrow class of
structures. This combination of features gives grounds to consider δ-concepts not just classification
indicators, but cognitive carriers of semantic integrity.</p>
        <p>In general, the author's model demonstrated a high level of generalisation and adaptation to the
data: the difference between the results on the validation and test samples was no more than 2.1%,
which indicates stable generalisation without signs of overtraining. Expert evaluation of 30 cases of
δ-explanations confirmed relevance in 26 out of 30 cases (86.7%), which confirms the practical
suitability of the model in explainable AI and cognitive semantics tasks.</p>
        <p>However, the proposed δ-relaxed explainable approach is not without certain limitations. The
main limitation is the assumption of the independence of the activity of δ-concepts in the attention
module, which, although it simplifies the interpretation, can lead to the loss of latent correlations
between formal features. Another significant limitation is the fixed value of the parameter δ in the
process of constructing the set of concepts. The invariance of this threshold limits the variability of
generalisation, especially in the conditions of mixed or unevenly balanced corpora. In addition,
classification experiments were conducted under the conditions of symmetric dichotomy of classes.
It ensured the stability of the metrics, but does not reflect realistic situations with uneven or
multimodal classes, typical of sociolinguistic or biomedical tasks. Finally, the current architecture of
the model does not take into account the reverse direction of interpretation - from explanation to
reconstruction of the input signal or generation of alternative solutions. It limits the application of
the model to counterfactual analysis, cognitive modelling, and neurointerface tasks. Despite the
factors listed, none of the limitations compromises the internal consistency of the model or its
explanatory nature. On the contrary, they outline a clear route for structurally extending the
architecture and increasing its adaptability in conditions of high semantic variability and latent
corpus heterogeneity.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusions</title>
      <p>The proposed δ-relaxed formal model for constructing explainable architectures is a relevant
response to the challenges of modern computational linguistics, especially in the context of increased
requirements for transparency, ethics, and trustworthiness of artificial intelligence in high-risk areas
- law, medicine, and education. Traditional binary formalisms of formal conceptual analysis turn out
to be insufficient for modelling language units with partially realised or context-dependent features,
which necessitates the need for a flexible δ-relaxed structure.</p>
      <p>The scientific novelty of the study is that for the first time, the full cycle of the δ-relaxed
explainable architecture has been not only theoretically substantiated, but also implemented - from
the formalisation of the conceptual grid to its software implementation, interpretation layer and
empirical verification. The proposed model differs from existing analogues (CBM, SHAP, LIME) by
the introduction of a multi-level hierarchy of δ-concepts, resistant to corpus variability, which allows
for automatic aggregation of significant features with subsequent interpretation of concepts in terms
of semantic features. The attention mechanism has been improved through the introduction of the
aggregated interest function  ( ,  ), which allowed building an explainable model with the
dominance of relevant concepts: the average Lift was 1.23 versus 1.04 in CBM, Δ-Stability – 0.77,
entropy  – 0.50. It provided a classification of sentences by pragmatic types (statements, directives,
questions, expressives) in the Ukrainian corpus with manually generated markup based on 128
concepts selected according to the criteria  &gt; 0.6 and  &lt; 0.65.</p>
      <p>The analysis of the experimental results confirmed the adequacy of the proposed model: the
δarchitecture achieved a macro-average F1-score of 0.84, precision is 0.86, and recall is 0.82, which
corresponds to the level of deep models without an explanatory layer. The difference between the
validation and test samples was no more than 2.1%, which demonstrates the ability of the model to
qualitative generalisation. According to the results of the Student's t-test (α = 0.05), the difference in
the Lift indicator between the δ-model and CBM is statistically significant (mean Lift: 1.23 vs. 1.04; p
= 0.049), while for the indicators Δ-Stability (p = 0.081) and the entropy of the target label Н (p =
0.11) an increase is observed that does not reach the threshold of statistical significance. The F1 score
for both models is not significantly different (p = 0.187), indicating that classification accuracy is
maintained. The attention heatmap in Figure 2 confirms the model's concentration on semantically
significant tokens, indicating interpretability of decision-making mechanisms.</p>
      <p>The practical value of the model lies in its ability to provide transparent classification with
formalised justification of results in applied NLP tasks: recognition of pragmatic functions of
statements in chatbots, legal examination of texts, cognitively guided educational systems,
interpreted recommendation modules. The interpretability of the model is realised not only through
weight coefficients but also through a structured ontology of concepts that are understandable for
both the researcher and the end user. However, the limitation is the need for pre-formation of
features based on manually created templates (e.g., semantic groups, grammatical dependencies),
which complicates application on multi-genre or multilingual corpora without adaptation. In
addition, the construction of δ-closures and the calculation of Lift and Δ functions for all pairs ( ,  )
has a complexity of  ( ), which imposes restrictions on the use of the model in real-time modes.</p>
      <p>Prospects for further research include automatic feature extraction from transformer models
(BERT, XLM-R), implementation of low-rank attention to reduce computational complexity,
extension of the δ-model to multimodal corpora (text + audio), as well as implementation of dynamic
concept formation in the inference process. Special attention is planned to be paid to the adaptation
of the model to multilingual environments, taking into account typological differences of languages
(morphological complexity, types of agreement) through flexible calibration of δ-thresholds.</p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgements</title>
      <p>The authors are grateful to all colleagues and institutions that contributed to the research and made
it possible to publish their results.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <sec id="sec-6-1">
        <title>The authors have not employed any Generative AI tools.</title>
        <p>[7] J. Li, G. Li, Y. Li, Z. Jin, Structured chain-of-thought prompting for code generation, ACM Trans.</p>
        <p>
          Softw. Eng. Methodol. 34 (
          <xref ref-type="bibr" rid="ref2">2</xref>
          ) (2025) 1–23. doi:10.1145/3690635.
[8] F. Rancourt, P. Vondrlik, D. Maupomé, M.-J. Meurs, Investigating self-rationalising models for
commonsense reasoning, Stats 6 (
          <xref ref-type="bibr" rid="ref3">3</xref>
          ) (2023) 907–919. doi:10.3390/stats6030056.
[9] S. Fatemi, Y. Hu, M. Mousavi, A comparative analysis of instruction fine-tuning large language
models for financial text classification, ACM Trans. Manag. Inf. Syst. 16 (
          <xref ref-type="bibr" rid="ref1">1</xref>
          ) (2025) 1–30.
doi:10.1145/3706119.
[10] A. Madsen, S. Reddy, S. Chandar, Post-hoc interpretability for neural NLP: A survey, ACM
        </p>
        <p>
          Comput. Surv. 55 (8) (2022) 1–42. doi:10.1145/3546577.
[11] L. Wu, A review of the transition from Shapley values and SHAP values to RGE, Statistics (2025)
1–23. doi:10.1080/02331888.2025.2487853.
[12] H. Moraliyage, G. Kulawardana, D. De Silva, Z. Issadeen, M. Manic, S. Katsura, Explainable
artificial intelligence with integrated gradients for the detection of adversarial attacks on text
classifiers, Appl. Syst. Innov. 8 (
          <xref ref-type="bibr" rid="ref1">1</xref>
          ) (2025) 17. doi:10.3390/asi8010017.
[13] P. Mehrani, J.K. Tsotsos, Self-attention in vision transformers performs perceptual grouping, not
attention, Front. Comput. Sci. 5 (2023). doi:10.3389/fcomp.2023.1178450.
[14] T. Babaian, J. Xu, Entity recognition from colloquial text, Decis. Support Syst. 179 (2024) 114172.
        </p>
        <p>doi:10.1016/j.dss.2024.114172.
[15] J. Yang, X. Hu, G. Xiao, Y. Shen, A survey of knowledge enhanced pre-trained language models,</p>
        <p>
          ACM Trans. Asian Low-Resour. Lang. Inf. Process. (2024). doi:10.1145/3631392.
[16] A.G. Vrahatis, K. Lazaros, S. Kotsiantis, Graph attention networks: A comprehensive review of
methods and applications, Future Internet 16 (9) (2024) 318. doi:10.3390/fi16090318.
[17] J.M. Rožanec, B. Fortuna, D. Mladenić, Knowledge graph-based rich and confidentiality
preserving explainable artificial intelligence (XAI), Inf. Fusion 81 (2022) 91–102.
doi:10.1016/j.inffus.2021.11.015.
[18] C. Zheng, Y. Wang, B. Chang, Query your model with definitions in FrameNet: An effective
method for frame semantic role labeling, in: Proc. AAAI Conf. Artif. Intell. 37 (11) (2023) 14029–
14037. doi:10.1609/aaai.v37i11.26642.
[19] Y.-R. Syau, E.-B. Lin, C.-J. Liau, An order-theoretic study on formal concept analysis, Axioms 12
(12) (2023) 1099. doi:10.3390/axioms12121099.
[20] L. Kovács, Concept lattice-based classification in NLP, in: Proc. 14th Int. Conf. Interdisciplinarity
in Engineering – INTER-ENG 2020, MDPI, 2020, p. 48. doi:10.3390/proceedings2020063048.
[21] F. Hao, J. Gao, C. Bisogni, G. Min, V. Loia, C. De Maio, Stability of three-way concepts and its
application to natural language generation, Pattern Recognit. Lett. 149 (2021) 51–58.
doi:10.1016/j.patrec.2021.06.005.
[22] O. Bisikalo, V. Kharchenko, V. Kovtun, I. Krak, S. Pavlov, Parameterisation of the stochastic
model for evaluating variable small data in the Shannon entropy basis, Entropy 25 (
          <xref ref-type="bibr" rid="ref2">2</xref>
          ) (2023)
184. doi:10.3390/e25020184.
[23] V. Kovtun, E. Zaitseva, V. Levashenko, K. Grochla, O. Kovtun, Small stochastic data
compactification concept justified in the entropy basis, Entropy 25 (12) (2023) 1567.
doi:10.3390/e25121567.
[24] V. Kovtun, V. Vysotska, O. Kovtun, The information-differentiated loss function for speech
feature clustering in a low-resource environment, in: CEUR Workshop Proc., vol. 3983, 2025, pp.
28–47. URL: https://ceur-ws.org/Vol-3983/paper3.pdf.
        </p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>N.</given-names>
            <surname>Balasubramaniam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kauppinen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rannisto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Hiekkanen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kujala</surname>
          </string-name>
          ,
          <article-title>Transparency and explainability of AI systems: From ethical guidelines to requirements</article-title>
          ,
          <source>Information and Software Technology</source>
          <volume>159</volume>
          (
          <year>2023</year>
          )
          <article-title>107197</article-title>
          . doi:
          <volume>10</volume>
          .1016/j.infsof.
          <year>2023</year>
          .
          <volume>107197</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>D.U.</given-names>
            <surname>Socol de la Osa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Remolina</surname>
          </string-name>
          ,
          <article-title>Artificial intelligence at the bench: Legal and ethical challenges of informing-or misinforming-judicial decision-making through generative AI</article-title>
          ,
          <source>Data &amp; Policy</source>
          <volume>6</volume>
          (
          <year>2024</year>
          ). doi:
          <volume>10</volume>
          .1017/dap.
          <year>2024</year>
          .
          <volume>53</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>D.E.</given-names>
            <surname>Mathew</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.U.</given-names>
            <surname>Ebem</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.C.</given-names>
            <surname>Ikegwu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.E.</given-names>
            <surname>Ukeoma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.F.</given-names>
            <surname>Dibiaezue</surname>
          </string-name>
          ,
          <article-title>Recent emerging techniques in explainable artificial intelligence to enhance the interpretable and understanding of AI models for human, Neural Process</article-title>
          .
          <source>Lett</source>
          .
          <volume>57</volume>
          (
          <issue>1</issue>
          ) (
          <year>2025</year>
          ).
          <source>doi:10.1007/s11063-025-11732-2.</source>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>C.M.</given-names>
            <surname>Greco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Tagarelli</surname>
          </string-name>
          ,
          <article-title>Bringing order into the realm of Transformer-based language models for artificial intelligence and law</article-title>
          ,
          <source>Artif. Intell. Law</source>
          <volume>32</volume>
          (
          <issue>4</issue>
          ) (
          <year>2023</year>
          )
          <fpage>863</fpage>
          -
          <lpage>1010</lpage>
          . doi:
          <volume>10</volume>
          .1007/s10506- 023-09374-7.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Duan</surname>
          </string-name>
          ,
          <article-title>A review of large language models: Fundamental architectures, key technological evolutions, interdisciplinary technologies integration, optimisation and compression techniques, applications, and challenges</article-title>
          ,
          <source>Electronics</source>
          <volume>13</volume>
          (
          <issue>24</issue>
          ) (
          <year>2024</year>
          )
          <article-title>5040</article-title>
          . doi:
          <volume>10</volume>
          .3390/electronics13245040.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>R.A.</given-names>
            <surname>Husein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Aburajouh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Catal</surname>
          </string-name>
          ,
          <article-title>Large language models for code completion: A systematic literature review</article-title>
          ,
          <source>Comput. Stand. Interfaces</source>
          <volume>92</volume>
          (
          <year>2025</year>
          )
          <article-title>103917</article-title>
          . doi:
          <volume>10</volume>
          .1016/j.csi.
          <year>2024</year>
          .
          <volume>103917</volume>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>