<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>V. Lytvyn, I. Peleshchak, R. Peleshchak, O. Mediakov, P. Pukach, Development of a hybrid
neural network model for mine detection by using ultrawideband radar data, Eastern-European
Journal of Enterprise Technologies</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="doi">10.15587/1729-4061.2023.279891</article-id>
      <title-group>
        <article-title>Detection and recognition of mines using magnetic field sensors and Kolmogorov-Arnold networks⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Roman Peleshchak</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vasyl Lytvyn</string-name>
          <email>vasyl.v.lytvyn@lpnu.ua</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ivan Peleshchak</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Viacheslav Beltiukov</string-name>
          <email>viacheslav.r.beltiukov@lpnu.ua</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Danylo</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Chystyakov</string-name>
          <email>chystyakov.d@nltu.lviv.ua</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yaroslav Kis</string-name>
          <email>yaroslav.p.kis@lpnu.ua</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Lviv Polytechnic National University</institution>
          ,
          <addr-line>12 Bandera Street, 79013, Lviv</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>National Forestry University of Ukraine</institution>
          ,
          <addr-line>103 Gen. Chuprynky St., 79057, Lviv</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>1990</year>
      </pub-date>
      <volume>3</volume>
      <issue>2023</issue>
      <fpage>0000</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>This paper presents a passive demining method that integrates fluxgate magnetic sensing with Kolmogorov-Arnold Networks (KANs) for multiclass landmine recognition. Magnetic-field anomalies recorded by an FLC-100 sensor above buried targets are combined with sensor height and categorical soil descriptors to form a three-dimensional feature space. To overcome data scarcity, 338 authentic measurements were augmented by injecting Gaussian noise that preserves subgroup statistics, enlarging each soil-mine pair by fifty samples and smoothing class distributions. Two spline-based architectures were compared: a compact KAN (3, 16, 16, 4) reaching accuracy of 93.56 %, and a wider KAN (3, 64, 64, 4) that offers further improvement to 95.59 % while virtually eliminating confusion between anti-personnel and booby-trap mines. Both models showed stable convergence without over-fitting, confirming the robustness of spline activations against sensor noise. Confusion-matrix analysis revealed perfect or nearperfect discrimination of “no-mine” and anti-tank cases, while remaining errors were localized to subtly differing magnetic signatures. The proposed detection is passive, avoiding the detonation risks associated with active probing and providing interpretable spline weights that expose feature contributions for safety certification. The results demonstrate the potential of physics-aware data augmentation and functional-edge neural architectures to accelerate safe demining operations.</p>
      </abstract>
      <kwd-group>
        <kwd>1Kolmogorov-Arnold network</kwd>
        <kwd>fluxgate magnetic sensor</kwd>
        <kwd>passive mine detection</kwd>
        <kwd>structured data</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Landmine detection remains a persistent and growing global concern, posing life-threatening risks
to millions of people. According to the Landmine Monitor 2023, landmines and explosive remnants
of war (ERW) continue to cause severe humanitarian consequences, with over 4,700 casualties
reported globally in 2022 alone, the vast majority of whom were civilians. More than 60 countries
remain contaminated by landmines, presenting ongoing risks for local populations, especially in
post-conflict regions such as Ukraine, where land access, agricultural activity, and reconstruction
are critically hindered [1]. In post-war Ukraine, the problem of landmine contamination has
become especially urgent, with vast areas of agricultural and residential land requiring safe
clearance. Traditional mine detection techniques often lack the reliability and responsiveness
needed for large-scale humanitarian demining. Moreover, many active detection methods—based
on emitting electric signals—risk triggering explosive devices, endangering human operators.</p>
      <p>A promising alternative is the use of passive detection systems [2], particularly those based on
magnetic field anomaly sensing [3, 4]. To enhance detection accuracy and reduce operational risks,
modern solutions increasingly rely on machine learning techniques, including neural networks.
However, neural networks often struggle with noise and distortions in real-world sensor data. One
method for improving the robustness and pattern recognition capability of neural architectures is
to use neural networks with embedded spline-based functional components, such as Kolmogorov–
Arnold Networks (KANs), which are particularly effective at handling noisy and irregular data due
to their ability to learn smooth, localized approximations of complex functions.</p>
      <p>The aim of this study is to develop an optimized architecture of Kolmogorov-Arnold Networks
in terms of the number of hidden layers, neurons per layer, and spline shape for accurate
recognition of mine types in soils of varying composition. This research is highly relevant in the
context of post-war recovery efforts in Ukraine, where effective and safe detection of minefields
plays a crucial role in restoring civil infrastructure and ensuring public safety.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Literature Review</title>
      <p>Kolmogorov–Arnold Networks (KAN) represent a recent advancement in neural network
architecture inspired by the Kolmogorov–Arnold representation theorem. Unlike traditional
multilayer perceptrons (MLPs) that rely on fixed activation functions at each node, KAN replaces
every weight with a univariate, spline-parametrized function. This allows KANs to learn richer
functional representations with fewer parameters, making them both efficient and interpretable
[5].</p>
      <p>In the work by Erdmann et al. [6], KAN was applied to a binary classification problem in
highenergy physics. The authors found that while multilayer KANs did not always outperform standard
MLPs in terms of accuracy, they demonstrated greater interpretability. Specifically, the activation
functions learned in deeper KANs differed significantly from those in shallow models, indicating
the architecture's capacity for more abstract feature extraction.</p>
      <p>Somvanshi et al. [7] provide a comprehensive survey on KAN, outlining its theoretical
foundations and practical adaptations across domains such as biomedical analytics, time series
prediction, and graph learning. They highlight KAN’s flexibility and adaptability, particularly in
handling high-dimensional structured data.</p>
      <p>Barasin et al. [8] explored KAN in the context of time series classification using the UCR
benchmark dataset. Their findings revealed that well-optimized KAN models outperformed MLPs
and achieved competitive results compared to state-of-the-art models such as HIVE-COTE2, all
while maintaining computational efficiency and robustness to hyperparameter changes.</p>
      <p>In terms of robustness, the study published in Applied Sciences assessed the vulnerability of
different KAN architectures to adversarial attacks [9]. Among the variants, KAN-Mixer showed the
best performance in resisting attacks while retaining strong accuracy on clean data. This makes
KAN suitable for safety-critical applications like mine detection, where robustness is paramount.</p>
      <p>In the field of remote sensing, Cheon [10] proposed combining pretrained CNNs with KAN
layers for scene classification using the EuroSAT dataset. The hybrid models achieved high
classification accuracy with reduced parameter counts and faster training, illustrating the potential
for integrating KAN into real-time systems.</p>
      <p>Drokin [11] extended KAN’s application to computer vision tasks, proposing
parameterefficient KAN convolution layers and fine-tuning techniques. The results demonstrated that
KANbased models can achieve strong performance in both image classification and segmentation tasks,
suggesting relevance to image-based mine detection scenarios.</p>
      <p>The reviewed literature suggests that KAN offers a unique combination of interpretability,
efficiency, and reliability across various classification domains. These characteristics make it a
promising candidate for mine detection, especially in post-war Ukraine, where safety,
dependability, and explainability are of paramount importance. However, the optimization of the
Kolmogorov-Arnold Network architecture to improve recognition accuracy – as well as the
tradeDefinition</p>
      <sec id="sec-2-1">
        <title>Limit values/Class off between training speed and recognition precision – remains an open challenge, which is crucial in the context of mine detection.</title>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Classification data</title>
      <p>In this study, we utilized the dataset provided in [4], which focuses on the classification of
landmines based on magnetic field anomaly characteristics. The parameter values employed in our
experimental setup are summarized in Table 1. Furthermore, we analyzed the relationship between
the magnetic anomaly values and the soil type (Table 2), as well as the distance between the
magnetic sensor and the buried landmine (Table 3). The general trends in magnetic field anomalies
across different landmine types were also examined and illustrated (Table 3).</p>
      <p>To obtain reliable measurements of the magnetic anomalies surrounding subsurface mines, the
original study [4] employed a fluxgate magnetic sensor model FLC100 [12], which demonstrated
sensitivity to minute variations in the magnetic field. This sensor-based approach enabled passive
mine detection without the need for active signal emission, thus reducing the risk of accidental
detonation. The design and deployment of the sensing mechanism were previously validated in [4],
where a decision support system for mine classification was developed using metaheuristic
classifiers.</p>
      <sec id="sec-3-1">
        <title>Voltage (V) High (H) Soil Type (S)</title>
        <p>aFocLutCitopTnsuehatonenvfsovoothmalrtleaduagmuleyeeoa.ogtfofnthtethheteiec dsiteshnteasnogcrrTeoahuobenfodtvh.ee tohfesso6tiadltidefefoeprfeemnndtoiitnsygtpuoersne. coclmalanTmsdyso;epn4selsoydfioffmoffeumirnneiednnseto.sn</p>
      </sec>
      <sec id="sec-3-2">
        <title>Output Data,</title>
        <p>“Dependent Variable”
Mine Type (M)
[0 V, 10.6 V]
[0 cm, 20 cm]</p>
      </sec>
      <sec id="sec-3-3">
        <title>Dry and sandy</title>
        <p>Dry and purulent
Dry and chalky
Wet and sandy
Humid and humus
Wet and chalky</p>
      </sec>
      <sec id="sec-3-4">
        <title>Null Anti-tank Anti-personnel</title>
      </sec>
      <sec id="sec-3-5">
        <title>Booby trapped Anti-personnel</title>
        <p>Soil Type Null, VAnti-Tank, VAnti-Personnel, VBooby Trapped Anti-Personnel, V
Dry and sandy 3.560 10.400 3.830 5.590
Dry and purulent 3.500 7.500 3.920 5.590
Dry and chalky 3.720 10.400 6.890 2.406
Wet and sandy 3.780 10.400 6.220 4.490
Humid and humus 3.350 10.400 5.050 2.770</p>
        <p>Wet and chalky 3.610 10.400 5.960 4.400</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Preprocessing data</title>
      <sec id="sec-4-1">
        <title>4.1. Data Generation Based on Parameterized Normal Distribution</title>
        <p>To improve the generalization capability of the model on a limited dataset consisting of 338 real
records [4], an additional data generation procedure was applied using a parameterized normal
distribution.</p>
        <p>The chosen method is based on generating new examples by adding pseudorandom noise [13]
to the original feature values V and H within each subgroup of data defined by a unique pair of
soil type S and mine type M . For each such subgroup, the statistical characteristics of the
features are computed as follows:</p>
        <p>V  V , V  std V  ,H  H , H  std  H  ,
V , H — mean values of features V and H , respectively.
 V , H — standard deviations of features V and H .</p>
        <p>V , H — arithmetic means of the respective columns.</p>
        <p>std  X  — standard deviation operator applied to feature X .</p>
        <p>New examples are generated using the following formulas:</p>
        <p>Vi'  Vi  N 0, V   , Hi'  H  N 0, H   ,</p>
        <p>i
where:</p>
        <p>
where:</p>
        <p>
sample.</p>
        <p>




</p>
        <p>Vi', Hi' — newly generated values of magnetic field anomaly and height for the  -th
Vi, Hi — values sampled from an existing record in the subgroup.
  0.1 — noise intensity coefficient (empirically selected).</p>
        <p>N 0,  — normally distributed random value with mean 0 and standard deviation
 .</p>
        <p>
          To ensure the physical plausibility of generated values, clipping was applied to constrain them
within real-world sensor bounds:
(
          <xref ref-type="bibr" rid="ref1">1</xref>
          )
(2)
in accordance with the sensor specifications.
        </p>
        <p>For each subgroup defined by  S, M  , 50 new samples were generated, which significantly
increased the number of training examples and smoothed the data distribution.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Data Preprocessing before training</title>
        <p>Before training the neural network, the following preprocessing steps were performed:
1. Normalization of magnetic anomaly feature V:</p>
        <p>V  V
V   ,</p>
        <p> V</p>
        <p>V  — normalized magnetic anomaly value.</p>
        <p>V — original value of the magnetic anomaly.</p>
        <p>V — mean magnetic anomaly over the entire dataset.
 V — standard deviation of the magnetic anomaly.
(3)
(4)
(5)
2. Categorical encoding of the soil type variable S, which takes six values:
 “Dry and Sandy”,
 “Dry and Humus”,
 “Dry and Limy”,
 “Humid and Sandy”,
 “Humid and Humus”,
 “Humid and Limy”.</p>
        <p>These categories were encoded using One-Hot Encoding, which transforms each category into
a binary vector of size six. For example, if the soil type is the second category (“Dry and Humus”),
the vector would be:</p>
        <p>0,1, 0, 0, 0, 0.</p>
        <p>
          3. Target encoding. The mine types M, originally in categorical form, were first
mapped to numerical indices (
          <xref ref-type="bibr" rid="ref1">0–3</xref>
          ) and then encoded using one-hot encoding for input into the
neural network.
        </p>
        <p>This method preserved the internal structure and semantics of the data, ensured physical
interpretability of the generated values, and significantly improved the model’s generalization
potential.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Mathematical Model of Kolmogorov–Arnold Networks</title>
      <sec id="sec-5-1">
        <title>5.1. Classification Problem Statement</title>
        <p>The articles [3, 4], addresses a multiclass-classification task solved with classical machine-learning
techniques — artificial neural networks [14] and their variants, support-vector machines [15],
Bayesian approaches [16], decision trees [17], and others.</p>
        <p>Let X be the feature space X = {V, H, S}, where V denotes the magnetic-field anomaly in the
vicinity of a mine (volts), H is the sensor height above the ground that covers the mine, and S
represents the soil type. The label set is Y = {0, 1, 2, 3}, whose elements correspond to the classes
“no mine,” “anti-tank mine,” “anti-personnel mine” and “booby-trap” respectively.</p>
        <p>The classification objective is to determine a mapping operator Y*: Х → Y that assigns any
previously unseen object x ∈ X to class y ∈ Y while minimizing the Euclidean error
min  y*  y ,
(6)
where y is the true class label and y* is the neural-network prediction.</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. KAN Morphology</title>
        <p>The Kolmogorov–Arnold Network (KAN) is a neural architecture inspired by the Kolmogorov–
Arnold representation theorem [18]. Unlike traditional MLPs that apply fixed nonlinearities at
nodes and learn linear weights, KANs apply learnable nonlinear activation functions on
edges, modeled as univariate splines. Each layer in a KAN consists of a matrix of spline
functions, and each neuron simply sums the outputs of these spline-parameterized edges.</p>
      </sec>
      <sec id="sec-5-3">
        <title>General Architecture</title>
        <p>We define a KAN with shape (3, 16, 16, 4) (Fig. 1), which consists of:
 Input layer with 3 nodes (features).
 Two hidden layers with 16 nodes each.</p>
        <p> Output layer with 4 nodes (classes or regression outputs).</p>
        <p>The general forward propagation is expressed by the composition of KAN layers:
KAN  x  Φ  Φ1  Φ0   x ,
2
(7)
where each Φl is a functional matrix consisting of learnable spline activations. Every layer
transforms its input by applying these univariate spline functions on each edge, followed by
summation at the next layer’s nodes.</p>
      </sec>
      <sec id="sec-5-4">
        <title>Second Hidden Layer (Layer 1 → 2)</title>
        <p>For each neuron k  1,,16 in the second hidden layer:</p>
      </sec>
      <sec id="sec-5-5">
        <title>Output Layer (Layer 2 → 3)</title>
        <p>For each output node m  1,, 4 :
xj1  i31φj0,i  xi0 .
x2  1j61φk1,j  x1 .</p>
        <p>k j
y3  16 φ2  x2 .</p>
        <p>m k 1 m,k k
The final output of the model is:
 </p>
        <p>3 
 y2
KAN  y    3   R
 y3 </p>
        <p>1
 y3 
 3 
 y4 
5.4. Spline Activation Functions
and a cubic B-spline [19]:</p>
      </sec>
      <sec id="sec-5-6">
        <title>5.3. Layer-wise Formulation</title>
        <p>follows:</p>
      </sec>
      <sec id="sec-5-7">
        <title>First Hidden Layer (Layer 0 → 1)</title>
        <p>For each neuron j  1,,16 in the first hidden layer:
Let the input vector be x0  x  R3 . The subsequent layer computations are defined as
(8)
(9)
(10)
(11)
(12)
Each edge activation function  jl,i  x is defined as a combination of a residual nonlinear term
φ  x   w  silu  x  w  
b s</p>
        <p>Gk1
m0
c B  x ,
m m






</p>
        <p>is the smooth SiLU function (acts as a residual base).</p>
        <p>Bm  x are cubic B-spline basis functions (order k  3 ).</p>
        <p>G  10 is the number of intervals → G  k  13 basis functions per spline.
cm are trainable spline coefficients.</p>
        <p>wb , ws are trainable scalar weights controlling the contribution of the SiLU and the spline.</p>
      </sec>
      <sec id="sec-5-8">
        <title>5.5. Parameter Count</title>
        <p>To compute the total number of parameters:</p>
        <p>Each spline has G  k  13 coefficients, plus 2 weights (w , w ) → 15 parameters per
b s
edge.</p>
        <p>First layer 0  : 316  48 edges → 48 15  720 parameters</p>
        <p>Second layer 1  :16 16  256 edges → 256 15  3840 parameters
Third layer  2  :16  4  64 edges → 64 15  960 parameters
Total parameters  720  3840  960  5520</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Computer experiment and results discussion</title>
      <sec id="sec-6-1">
        <title>6.1. Computer experiment</title>
        <p>In this section, we present the setup and execution of the computer experiment aimed at evaluating
the performance of Kolmogorov–Arnold Networks (KANs) for a multiclass classification task. The
experiment involved training and comparing two network architectures: KAN (3, 16, 16, 4) and
KAN (3, 64, 64, 4). The architectures were chosen based on the task's dimensionality, where the
input space was three-dimensional, and the output space consisted of four distinct classes.</p>
        <p>The training was conducted using the PyKAN library [20] on the Python platform. The settings
for the networks included the use of cubic B-splines as basis functions, with the order set to 3 and
the grid size set to 10. These parameters provided a sufficient balance between the flexibility of the
spline approximation and computational efficiency.</p>
        <p>The network KAN (3, 16, 16, 4) was trained first. It underwent a training process over a 65
epochs and achieved an accuracy of 93.56% on the test set. In the second case, the KAN (3, 64, 64, 4)
architecture was trained. Due to its significantly higher number of parameters, the training took a
considerably longer time; however, it achieved an improved test accuracy of approximately 95,59%.</p>
        <p>During the training process, loss and accuracy curves were recorded for each model to monitor
convergence dynamics and to detect potential signs of overfitting. After the evaluation phase,
confusion matrices were generated to provide a detailed understanding of classification
performance across all classes. In addition to the visual analyses, a comprehensive classification
report was produced, presenting key metrics such as precision, recall, and F1-score for each class.</p>
        <p>The details of the experimental environment, including the software tools and libraries, are as
follows:</p>
        <p>Programming language: Python
Neural network library: PyKAN [20]
Hardware: Personal PC (AMD Ryzen 5 5600G CPU, NVIDIA GeForce RTX 4060 GPU)
Software:
 IntelliJ IDEA (with Python plugin support).
 Python 3.x.
 PyKAN library [20].
 CUDA Toolkit (for GPU acceleration with NVIDIA RTX 4060).
 PyTorch (backend library for PyKAN).
 NumPy (for data manipulation).</p>
        <p> Matplotlib (for visualization of results).</p>
      </sec>
      <sec id="sec-6-2">
        <title>6.2. Discussion of Results</title>
        <p>The results of the experiments are illustrated through loss-accuracy curves, confusion matrices,
and a set of other performance metrics [21], which comprehensively describe the behavior of both
tested architectures.</p>
        <p>For the KAN (3, 16, 16, 4) network, the loss curve (Figure 2) demonstrated a steady decrease
without abrupt oscillations, indicating stable convergence. The corresponding accuracy curve
(Figure 2) showed consistent improvement throughout the training process, reaching a plateau
near 92.56%. The confusion matrix (Figure 3 and Figure 4) revealed that most misclassifications
occurred between the (specify which classes if possible), suggesting that the network found these
classes harder to differentiate given the feature space.</p>
        <p>The normalized confusion matrix for the KAN (3, 16, 16, 4) model, shown in Figs. 3 and 4,
reflects almost perfect identification of the "no mine" and "anti-tank" classes, with correct detection
rates reaching approximately ninety-seven percent. At the same time, the majority of
misclassifications occurred between the "anti-personnel" and "booby trap" classes: around ten
objects from the first category were confused with the second, while the reverse misclassification
happened almost twice as rarely. This asymmetry is explained by the partial overlap of magnetic
anomaly ranges and sensor height, indicating that the three-dimensional feature space was
insufficient to fully separate these mine types.</p>
        <p>Despite this, the model exhibits stable convergence of the loss function and absence of sharp
fluctuations, indicating proper hyperparameter tuning and sufficient capacity for the basic task.
However, it also signals the need to enrich the feature space specifically in the area where
classification errors are observed.
Precision - The proportion of predicted positive samples that are actually correct for each
class.</p>
        <p>Recall - The proportion of actual positive samples that are correctly predicted for each
class.</p>
        <p>F1-score - The harmonic mean of precision and recall for each class, providing a balance
between the two metrics.</p>
        <p>Support - The number of true instances for each class in the test set.</p>
        <p>Precision - The proportion of predicted positive samples that are actually correct for each
class.
0, 1, 2, 3 - The performance metrics for each individual class.</p>
        <p>Accuracy - The overall classification accuracy across all classes (i.e., the proportion of
correctly classified samples).</p>
        <p>Macro avg - The unweighted mean of precision, recall, and F1-score across all classes,
treating each class equally regardless of its support.</p>
        <p>Weighted avg - The mean of precision, recall, and F1-score weighted by the number of
true instances (support) for each class, giving more influence to classes with more samples.</p>
        <p>On the other hand, the KAN (3, 64, 64, 4) architecture, while requiring longer training time due
to the increased number of neurons, achieved superior classification results with approximately
95% accuracy. Its loss curve (Figure 5) exhibited a smoother descent, and its accuracy curve (Figure
5) achieved a slightly higher and more stable plateau compared to the smaller network. The
confusion matrix (Figure 6 and Figure 7) for this model showed a significant reduction in
misclassification rates across all classes, particularly improving recognition of (specify if needed).</p>
        <p>Increasing the number of neurons to sixty-four in each hidden layer led to significant changes
in the error patterns, as clearly seen in the confusion matrices in Figs. 6 and 7. The updated KAN
(3, 64, 64, 4) architecture almost completely eliminated confusion between the "anti-personnel" and
"booby trap" classes in the direction from the latter to the former, raising the accuracy for the
"booby trap" class above ninety-five percent. Reverse confusion still occurred in about seven cases
out of seventy-three, reducing the recall of this class to ninety-four percent, but these mistakes
now have a one-sided nature: the network becomes more conservative, assigning doubtful samples
to the less dangerous category in the absence of convincing evidence. The increased computational
costs are justified by the fact that overall classification accuracy improved by about two percent,
and the off-diagonal elements of the matrix sharply decreased for all classes except for the localized
issue of booby trap identification.</p>
        <p>The comparison of the two models shows that even with the same basic set of features, a wider
architecture can capture finer signal nonlinearities and thus reduce the number of critical errors.
At the same time, the remaining confusion between classes 2 and 3 indicates the limit beyond
which pure network scaling becomes less effective compared to introducing additional
information, such as gradient characteristics of the magnetic field or contextual soil indicators.</p>
        <p>Thus, detailed analysis of the confusion matrices indicates that the main direction for further
optimization should be strengthening the discriminative power of features specifically for the
"booby trap" class, while preserving the already achieved high reliability in detecting other mines
and safe areas.
Weighted avg</p>
        <p>Comparative analysis of the two models (Table 4 and Table 5) indicates that increasing the
hidden layer size improves generalization capability but at the cost of greater computational time
and resources. This trade-off must be considered depending on the application domain
requirements.</p>
        <p>Thus, the computer experiments have demonstrated that Kolmogorov–Arnold Networks, when
properly configured with cubic B-splines and an appropriate grid resolution, can achieve high
accuracy in multiclass classification problems, with performance scaling positively with network
capacity.</p>
        <p>Study Limitations</p>
        <p>The base dataset has a limited volume; although synthetic augmentation improves
generalization, it cannot fully replace field measurements. The results were obtained under
laboratory conditions without considering the influence of metallic debris, heterogeneous magnetic
backgrounds, or sensor temperature drifts.</p>
        <p>Future Research Directions</p>
        <p>Collection of large-scale field data under various climatic and geological conditions to validate
the results.</p>
        <p>End-to-end optimization: selection of spline grids, nonlinearity bases, and regularization
techniques (e.g., KAN-Mixer, weight priorities) to further improve accuracy without exponential
growth in parameters.</p>
        <p>Robustness: investigation of resilience to adversarial influences typical of deceptive mine
masking with metallic shrapnel or geomagnetic traps.</p>
        <p>7. Conclusions
1. For the first time, Kolmogorov–Arnold Networks (KAN) with cubic B-splines were used for
passive mine recognition based on magnetic anomalies. Unlike classical MLPs, KAN allows
modeling nonlinear dependencies at the level of weight connections, enhancing the interpretability
and robustness of the model to noise in real sensor measurements.</p>
        <p>2. An extended dataset was created: synthetic samples were added to 338 original magnetic
anomaly recordings, generated using a parameterized normal distribution with a step of 50 samples
for each "soil type – mine type" subset. This balanced the feature variance and reduced the risk of
overfitting.</p>
        <p>3. Two architectures were developed: KAN (3, 16, 16, 4) and KAN (3, 64, 64, 4). Both models
were trained using the PyTorch-compatible PyKAN library with identical spline hyperparameters.</p>
        <p>KAN (3, 16, 16, 4) achieved 93.56% accuracy without signs of overfitting; the main errors
occurred between the "anti-personnel" and "booby trap" classes. Increasing the number of neurons
in KAN (3, 64, 64, 4) to 64 per hidden layer improved accuracy to 95.59%, significantly reducing
false detections across all four classes. The cost of this improvement was an almost linear increase
in the number of parameters and training time. This confirms the advisability of adaptively
selecting model size based on the hardware constraints of field systems. Confusion-matrix analysis
showed both models nearly flawless at identifying “no-mine” and anti-tank cases, while most
errors arose from confusion between anti-personnel mines and booby-traps. Results confirm that
increasing network capacity improves discrimination among visually similar magnetic signatures.
4. Advantages of the proposed approach.</p>
        <p>Passive mine recognition: the use of the FLC-100 sensor does not require active excitation,
minimizing the risk of detonation.</p>
        <p>Interpretability: spline weights enable analysis of the contribution of each feature and facilitate
safety certification.</p>
        <p>Robustness: experiments showed no sharp fluctuations in the loss function and stable
convergence even on a noise-enriched dataset.</p>
        <p>The study proves that properly configured Kolmogorov–Arnold Networks can achieve over 95%
accuracy in multiclass mine classification based on passive magnetic features. The combination of
interpretable spline weights, high sensitivity to small anomalies, and scalability potential makes
KAN a promising foundation for modular humanitarian demining systems, which can significantly
accelerate land clearance and reduce risks for personnel.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgements</title>
      <p>The research was carried out with the grant support of the National Research Fund of Ukraine
"Methods and means of active and passive recognition of mines based on deep neural networks",
project registration number 273/0024 from 1/08/2024 (2023.04/0024). Also, we would like to thank
the reviewers for their precise and concise recommendations that improved the presentation of the
results obtained.</p>
    </sec>
    <sec id="sec-8">
      <title>Declaration on Generative AI</title>
      <p>The authors have not employed any Generative AI tools.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <article-title>[1] Landmine and Cluster Munition Monitor</article-title>
          ,
          <source>Landmine Monitor</source>
          <year>2023</year>
          , International Campaign to Ban Landmines - Cluster
          <string-name>
            <surname>Munition Coalition (ICBL-CMC)</surname>
          </string-name>
          (
          <year>2023</year>
          ). URL: https://backend.icblcmc.org/assets/reports/LandmineMonitors/LMM2023/Downloads/Landmine-Monitor-2023_web.pdf
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>