<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>O. Bisikalo);</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>systems⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Oleg Bisikalo</string-name>
          <email>obisikalo@vntu.edu.ua</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mariya Yukhymchuk</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pavlo Strembitskyi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vladyslav Lesko</string-name>
          <email>Leskovlad@ukr.net</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Vinnytsia National Technical University</institution>
          ,
          <addr-line>Khmelnytske Shosse St. 95 21021 Vinnytsia</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Vinnytsia</institution>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>Law enforcement equipment failures during critical operations create immediate risks to officer safety and public security. Traditional monitoring relies on reactive maintenance - fixing problems after they occur. This paper presents a monitoring framework that combines Prometheus and Thanos infrastructure with three machine learning techniques: adaptive Isolation Forest for anomaly detection, LSTM networks with attention mechanisms for failure prediction, and reinforcement learning for parameter optimization. The system was validated through comprehensive cloud-based simulation modeling six months of operations across 1,847 synthetic monitoring points representing radio systems, surveillance cameras, and vehicle equipment. Simulation parameters were derived from published equipment failure statistics and validated through limited pilot deployment (87 endpoints, 3 months) with one partner law enforcement agency. Results show F1-score improvement from 0.72 to 0.89, detection time reduction from 147 to 41 seconds (72% faster), and false positives dropping from 12.3% to 3.8%. LSTM models predicted equipment failures 4-8 hours in advance with 87% average accuracy across five equipment categories. The framework scaled linearly from 50 to 3,000+ endpoints with detection latency under 52ms. However, several challenges remain: simulation cannot capture all real-world complexity, integration requires custom interfaces for legacy systems, and operators need enhanced explainability features to trust AI recommendations. While primary findings are based on simulated data, results suggest promising directions for operational systems pending comprehensive real-world validation.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;security monitoring</kwd>
        <kwd>law enforcement systems</kwd>
        <kwd>machine learning</kwd>
        <kwd>anomaly detection</kwd>
        <kwd>predictive maintenance</kwd>
        <kwd>LSTM networks</kwd>
        <kwd>explainable AI 1</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Equipment reliability fundamentally impacts law enforcement operational safety. Consider an
officer pursuing a suspect through an unfamiliar neighborhood when the patrol vehicle's radio
system fails. Without communication capability, the officer cannot request backup, cannot
coordinate with other units, and has no way to report current location or situation developments.
The pursuit continues but now with substantially elevated risk to both the officer and public safety.
This type of equipment failure occurs more frequently than most people realize, and the
consequences range from operational inefficiency to genuine danger for officers and communities
they serve.</p>
      <p>
        Documentation from the International Association of Chiefs of Police indicates that equipment
failures during operations create cascading effects extending far beyond immediate technical
problems [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. A malfunctioning body camera means lost evidence that could prove critical in
prosecution or exoneration.
investigations.
      </p>
      <p>A failing surveillance system might allow suspects to escape detection during time-sensitive</p>
      <p>Vehicle-mounted computer systems that crash during traffic stops leave officers without access
to warrant information, wanted person alerts, or backup unit locations. Each failure represents not
just broken equipment but a potential safety hazard and a gap in operational capability that might
prove consequential.</p>
      <p>
        Traditional approaches to law enforcement equipment monitoring reflect reactive maintenance
strategies common across many organizations. Agencies typically address equipment problems
only after obvious signs of failure appear. Routine maintenance follows fixed schedules based on
manufacturer recommendations rather than actual equipment condition. Threshold-based alerting
systems generate warnings only after problems have already begun manifesting in observable
symptoms. This reactive posture made sense historically when equipment was relatively simple
and agencies operated with limited resources for sophisticated monitoring infrastructure. However,
modern law enforcement technology has grown substantially more complex while the implications
of equipment failure have increased correspondingly [
        <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
        ].
      </p>
      <p>
        The contrast with monitoring capabilities in modern data center operations proves instructive.
Large technology companies have developed systems that predict hardware failures days in
advance with accuracy exceeding ninety percent [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Their monitoring infrastructure analyzes
thousands of metrics per second, detects subtle patterns that precede failures, and automatically
triggers preventive maintenance before problems impact operations. These capabilities exist
because data centers operate in controlled environments with consistent workloads, extensive
historical data, and substantial resources dedicated to infrastructure management.
      </p>
      <p>Law enforcement equipment faces fundamentally different challenges that complicate direct
technology transfer from data center contexts. Law enforcement equipment operates under
conditions that challenge standard monitoring approaches in multiple ways. A patrol vehicle might
sit idle in a parking lot for several hours, then operate continuously during a major incident
spanning an entire shift. Body cameras experience environments ranging from climate-controlled
offices through extreme heat, severe cold, heavy rain, and occasional physical impacts from
operational activities. Radio systems must function reliably whether officers work inside concrete
buildings with poor signal penetration, in rural areas with limited infrastructure, or in urban
environments with substantial electromagnetic interference. Network equipment serves users who
move constantly rather than remaining at fixed workstations with predictable connectivity
patterns. This operational variability means that patterns considered normal in one context might
indicate serious problems in another, complicating the design of monitoring systems that must
distinguish between legitimate operational variations and genuine equipment malfunctions.</p>
      <p>This research addresses three fundamental questions about applying AI-enhanced monitoring to
law enforcement equipment reliability. First, can machine learning techniques deliver performance
improvements substantial enough to justify the implementation complexity, computational costs,
and organizational changes required for deployment? Second, what practical obstacles arise when
deploying AI systems in security-sensitive environments with strict access controls, audit
requirements, and compliance obligations? Third, does the monitoring approach scale effectively
from small rural agencies managing dozens of devices through large metropolitan departments
operating thousands of endpoints across multiple facilities and organizational divisions?</p>
      <p>
        Rather than attempting to replace existing monitoring infrastructure with entirely new systems,
this work augments proven open-source tools with custom AI components designed specifically for
law enforcement operational requirements. The foundation consists of Prometheus for metrics
collection and Thanos for federated storage and cross-site querying [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. These established
technologies handle the infrastructure concerns of gathering metrics at scale, storing them reliably,
and providing efficient query capabilities. Three AI components layer on top of this foundation to
add intelligence that traditional threshold-based monitoring cannot provide.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Related works</title>
      <p>The application of machine learning to security equipment monitoring builds on research across
multiple domains including manufacturing quality control, infrastructure monitoring, time series
analysis, and explainable artificial intelligence. However, relatively few studies address the specific
challenges of operational equipment monitoring in law enforcement contexts where usage patterns
are highly variable, security requirements are stringent, and failure consequences can prove severe.</p>
      <p>
        Zhang and colleagues conducted extensive research on quality control applications in defense
manufacturing facilities where subtle defects in components can compromise equipment
performance during critical operations [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Their work demonstrated convincingly that traditional
statistical process control methods often miss anomalies that machine learning algorithms
successfully detect through pattern recognition in high-dimensional measurement spaces.
However, their research focused on manufacturing processes rather than operational monitoring of
deployed equipment in field environments.
      </p>
      <p>
        Kumar and Singh performed a comprehensive systematic review of predictive maintenance
approaches for security equipment, synthesizing findings from numerous studies across industrial
and defense applications [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Their analysis revealed a consistent pattern where research validated
on equipment with predictable usage patterns and controlled operating conditions frequently failed
to translate effectively to operational environments characterized by high variability and
unpredictable stress.
      </p>
      <p>
        The evolution of distributed monitoring infrastructure has tracked the increasing scale and
complexity of modern systems. Prometheus emerged as a de facto standard for metrics collection in
containerized and cloud-native environments. However, Prometheus was originally designed for
monitoring individual clusters rather than federating data across geographically distributed sites.
Thanos addresses these limitations by adding capabilities for long-term storage, global querying,
and high availability [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Strembitskyi and colleagues explored practical Thanos adoption
challenges in production deployments, documenting both successes and obstacles encountered
during real-world implementations.
      </p>
      <p>
        Yukhymchuk and colleagues investigated remote monitoring approaches for improving quality
and operational efficiency in production environments [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. They documented data quality
challenges that directly parallel issues encountered in law enforcement equipment monitoring,
including sensor drift, network connectivity problems, measurement noise, and time
synchronization issues. Their preprocessing techniques informed the data pipeline design described
in Section 3.
      </p>
      <p>
        Recent advances in observability engineering emphasize the critical importance of correlating
multiple telemetry types rather than relying exclusively on metrics alone [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Sridharan's work
demonstrated that combining metrics, logs, and traces provides substantially richer context for
understanding system behavior than any single data source achieves independently.
      </p>
      <p>
        Anomaly detection research has produced numerous algorithms suitable for different data
types. Isolation Forest has gained substantial popularity for its computational efficiency on
highdimensional data [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. However, standard implementations treat all detected anomalies as equally
deserving of attention, which generates excessive false alarms in environments where operational
context matters.
      </p>
      <p>
        Rodriguez and colleagues applied Soft Actor-Critic reinforcement learning to manufacturing
process optimization [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. Their implementation showed particular promise but they noted
important concerns about safety and reliability when applying RL to critical systems where
exploration could cause temporary performance degradation.
      </p>
      <p>
        García and colleagues performed a comprehensive survey cataloging data quality challenges
that affect time series analysis across diverse domains [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. They identified missing data, irregular
sampling, sensor drift, measurement noise, and outliers as common problems that appear
frequently in operational monitoring.
      </p>
      <p>
        Chen and Liu developed interpretable anomaly detection methods specifically designed for
critical infrastructure monitoring [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Their approach demonstrated that explainability and
prediction performance need not be mutually exclusive even for complex machine learning models.
Law enforcement applications face particularly stringent explainability requirements where
operators must understand system decisions to make informed responses.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <sec id="sec-3-1">
        <title>3.1. Simulation environment design</title>
        <p>
          The AWS-based simulation environment models multi-site law enforcement deployments spanning
four geographic regions representing different organizational contexts [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. While this represents
synthetic data rather than actual law enforcement operations, the simulation design incorporates
equipment failure rates from published law enforcement technology reports [
          <xref ref-type="bibr" rid="ref1 ref3 ref4">1, 3, 4</xref>
          ], operational
patterns described in academic literature [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ], and manufacturer specifications for metric ranges and
failure modes. This methodology enables controlled experimentation and reproducible research
while addressing realistic equipment characteristics that would be difficult to study in operational
environments due to security constraints and data availability limitations.
        </p>
        <p>Each region runs multiple EC2 instances generating realistic metrics using custom exporters
that model actual equipment behavior patterns. The overall architecture follows the Thanos-based
design illustrated in Figure 1, showing how Prometheus instances at each site connect through
Thanos Sidecars to centralized object storage, enabling global querying and long-term retention
across distributed monitoring points.</p>
        <p>
          The simulation generates approximately 45,000 metrics per second across all regions. Failure
injection follows chaos engineering principles adapted from Netflix's work on system resilience
testing [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ]. Rather than randomly introducing failures, the simulation schedules them at times and
in patterns that reflect how equipment actually fails in operational contexts based on published
failure statistics. The six-month simulation incorporated 847 distinct failure events distributed
across equipment categories matching documented reliability data.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. AI component architecture</title>
        <p>
          The monitoring architecture deliberately builds on established open-source tools rather than
creating proprietary systems from scratch. Prometheus handles metrics collection while Thanos
provides long-term storage and global querying capabilities [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. Three AI components layer on top
of this foundation.
        </p>
        <p>
          The first component adapts Isolation Forest for law enforcement operational contexts through
contextual weighting that adjusts anomaly scores based on current conditions [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]. Standard
Isolation Forest treats all anomalies identically, but the modification incorporates security and
quality factors that reflect operational context including threat levels and equipment criticality.
        </p>
        <p>The second component employs LSTM networks with attention mechanisms for failure
prediction. The architecture consists of three LSTM layers with dropout regularization to prevent
overfitting. The attention mechanism enables selective focus on relevant historical patterns, with
domain-specific importance weighting beyond standard learned attention scores.</p>
        <p>
          The third component uses Soft Actor-Critic reinforcement learning for automatic parameter
tuning [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]. The agent operates in an environment where states represent current monitoring
performance and actions represent parameter adjustments. Conservative exploration parameters
ensure the agent avoids dangerous experimentation that might compromise monitoring
effectiveness.
        </p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Mathematical models and algorithms</title>
        <p>
          The integrated monitoring framework builds upon the Prometheus and Thanos ecosystem while
adding custom AI components addressing law enforcement requirements [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]. The architecture
consists of six layers: data collection, storage federation, AI analytics, decision support, response
automation, and security compliance. Anomaly Detection: The research modified the standard
Isolation Forest algorithm for law enforcement characteristics. The standard anomaly score:
(1)
(2)
−E(h(x))
s ( x , n)=2 c(n)
where E(h(x)) represents average path length for isolating point x, and
c ( n)=2 ∙ H ∙ ( n−1)− 2 ∙ ( n−1) is the average path length of unsuccessful search in a Binary
n
Search Tree.
        </p>
        <p>Our law enforcement adaptation adds contextual weighting:</p>
        <p>slaw ( x , n)=2(− E(h(x))/claw(n))× wsecurity × wquality
where wsecurity and wquality are dynamic weighting factors adjusting based on current security threat
levels and equipment criticality scores.</p>
        <p>Predictive Maintenance: LSTM networks with attention mechanisms forecast equipment
failures. The attention mechanism calculates context vector ct as:</p>
        <p>T
ct=∑ alphati×hi×importancelaw ,i
i=1
(3)
where importancelaw,i weights historical periods based on operational context including mission
criticality, environmental conditions, and equipment stress levels.</p>
        <p>Process Optimization: Reinforcement learning using Soft Actor-Critic automatically adjusts
monitoring parameters. The reward function balances multiple law enforcement objectives:
R ( s , a)=w1×Rquality+w2×Rsecurity+w3×Rcompliance+w4×Refficiency
(4)
where weight parameters dynamically adjust based on operational priorities.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Data processing pipeline</title>
        <p>
          Raw metrics from Prometheus require extensive preprocessing before machine learning analysis.
The pipeline addresses missing values, sensor drift, measurement noise, outliers, and time
synchronization problems following approaches informed by prior work on data quality in time
series [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ].
        </p>
        <p>Feature engineering transforms raw metrics into derived features including rolling statistics,
rate of change measures, cross-correlations, and time series decomposition. Domain-specific
features incorporate equipment knowledge about patterns indicating developing problems.</p>
      </sec>
      <sec id="sec-3-5">
        <title>3.5. Model training strategy</title>
        <p>
          Training uses temporal splitting where models train on older data and test on newer data,
mimicking operational deployment where future events are predicted. Cross-validation employs
rolling forecast methodology. Hyperparameter optimization used grid search over reasonable
parameter ranges. Ensemble methods combine multiple models for robustness [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]. Model
calibration through temperature scaling improves probability estimates. Training leveraged
distributed GPU computing on AWS p3.2xlarge instances.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments</title>
      <sec id="sec-4-1">
        <title>4.1. Infrastructure configuration</title>
        <p>
          The experimental infrastructure used AWS m5.2xlarge instances (8 vCPUs, 32 GB RAM) for
monitoring components and p3.2xlarge instances with V100 GPUs for model training. Storage
utilized S3 with Intelligent-Tiering following AWS best practices [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. Software versions included
Prometheus 2.40.0, Thanos 0.30.0, Python 3.10.8, TensorFlow 2.12.0, PyTorch 1.13.1, and
Scikitlearn 1.2.0.
        </p>
        <p>The simulation generated metrics from 1,847 synthetic monitoring endpoints across five
equipment categories: radio communication systems (412 endpoints), surveillance camera networks
(563 endpoints), mobile computing devices (298 endpoints), network infrastructure equipment (387
endpoints), and vehicle-mounted systems (187 endpoints). Each endpoint generated between 15...40
metrics per minute depending on equipment type. The total dataset comprised approximately 2.4
billion individual metric measurements over the six-month evaluation period.</p>
        <p>
          Equipment failure rates and patterns were based on published statistics from law enforcement
technology reports [
          <xref ref-type="bibr" rid="ref1 ref3 ref4">1, 3, 4</xref>
          ]. For example, radio systems exhibited failure rates of 3-4% annually
with common failure modes including battery degradation, antenna damage, and software glitches.
Surveillance cameras showed higher failure rates (4-5% annually) with environmental factors and
mechanical wear being primary causes. The simulation incorporated these realistic failure
characteristics while maintaining complete ground truth annotations for rigorous evaluation.
        </p>
        <p>
          The baseline comparison system implemented sophisticated threshold-based monitoring with
static thresholds, rate-of-change rules, and correlation rules representing current best practices as
described in law enforcement technology literature [
          <xref ref-type="bibr" rid="ref1 ref4">1, 4</xref>
          ]. This baseline provides realistic
comparison rather than comparing against obviously inadequate systems.
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Experimental protocol</title>
        <p>The six-month evaluation divided into three phases. Phase 1 (Months 1-2) focused on initial
training and validation, with models trained on the first six weeks and validated on weeks 7-8.
Phase 2 (Months 3-5) simulated realistic operational deployment with monthly retraining cycles to
evaluate adaptation to evolving patterns. Phase 3 (Month 6) conducted stress testing with doubled
failure injection rates and edge case scenarios to evaluate robustness under challenging conditions.</p>
        <p>Evaluation employed comprehensive metrics including precision, recall, F1-score for detection
performance; detection latency and mean time to alert for temporal performance; prediction
accuracy and mean absolute error for predictive maintenance; false positive rate and resource
utilization for operational overhead; and throughput with latency under load for scalability
assessment.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Implementation details</title>
        <p>The Isolation Forest used 200 trees with adaptive contamination parameters (0.03-0.06) varying by
equipment category based on historical failure rates. Feature engineering expanded raw metrics
(~20 per endpoint) to ~100 features including rolling statistics, rate of change, and cross-equipment
comparisons.</p>
        <p>LSTM architecture consisted of three layers (128, 64, 32 units) with dropout (0.3) and attention
mechanism for selective focus on relevant historical patterns. Models processed 24-hour input
sequences to predict failure probability over the next 8 hours. Training used Adam optimizer
(learning rate 0.001), batch size 64, early stopping, and data augmentation.</p>
        <p>Reinforcement learning employed Soft Actor-Critic with state representation including current
performance metrics, alert volume, and resource utilization. Actions represented parameter
adjustments within safe ranges. The reward function balanced detection rate, false positive
reduction, response speed, and resource usage with weights reflecting operational priorities.</p>
        <p>Training leveraged distributed computing with data parallelism across 4 GPUs, achieving
approximately 3× speedup. All models were saved with complete versioning information enabling
reproducibility.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Results</title>
      <sec id="sec-5-1">
        <title>5.1. Anomaly detection performance</title>
        <p>Six-month evaluation demonstrates significant improvements over baseline monitoring
approaches. The modified Isolation Forest algorithm achieved F1-score of 0.89 versus 0.72 for
traditional threshold-based alerting, with false positive reduction from 12.3% to 3.8%.</p>
        <sec id="sec-5-1-1">
          <title>Detection Time (seconds)</title>
        </sec>
        <sec id="sec-5-1-2">
          <title>False Positive Rate</title>
        </sec>
        <sec id="sec-5-1-3">
          <title>Mean Time to Alert</title>
          <p>Response time improvements were particularly notable, with AI systems detecting simulated
equipment anomalies in average 41 seconds compared to 147 seconds for traditional approaches.
This improvement could provide valuable additional time for operators to address developing
problems before they impact operations. The system demonstrated consistent performance across
different equipment types and operational scenarios. Performance degradation in noisy
environments was minimal, with F1-scores remaining above 0.85 even with 20% measurement
noise.</p>
        </sec>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Predictive maintenance performance</title>
        <p>Predictive maintenance showed mixed results highlighting both potential and limitations. For
simulated communication equipment, LSTM models achieved 91% accuracy predicting failures 4...8
hours in advance. Mean absolute error in failure time prediction was 5.4 hours for radio systems.</p>
        <p>Equipment Type Prediction Accuracy
Lead Time (hours)</p>
        <sec id="sec-5-2-1">
          <title>Surveillance Cameras</title>
        </sec>
        <sec id="sec-5-2-2">
          <title>Mobile Devices</title>
        </sec>
        <sec id="sec-5-2-3">
          <title>Network Equipment</title>
        </sec>
        <sec id="sec-5-2-4">
          <title>Vehicle Systems</title>
          <p>Overall Average</p>
          <p>The attention mechanism proved valuable for handling irregular operational patterns typical of
law enforcement equipment. Traditional predictive models often assume regular usage patterns,
but law enforcement equipment may be heavily used during major events and idle for extended
periods.</p>
        </sec>
      </sec>
      <sec id="sec-5-3">
        <title>5.3. Scalability analysis</title>
        <p>The system scaled linearly from small agency configurations with 50 endpoints through large
metropolitan deployments with over 3,000 endpoints.
52ms
120K metrics/s</p>
        <p>Detection latency remained under 52 milliseconds even at the largest scale. Accuracy improved
with scale due to more training data, from 0.87 at small scale to 0.91 at metropolitan scale.
Throughput scaled from 5,000 to 120,000 metrics per second, substantially exceeding typical
requirements and providing headroom for growth.</p>
      </sec>
      <sec id="sec-5-4">
        <title>5.4. Real-world pilot testing observations</title>
        <p>To validate simulation realism and assess practical deployment challenges, a limited pilot
deployment was conducted with one partner law enforcement agency over three months. This pilot
provided the only real operational data in this research and revealed important insights about the
gap between simulation and reality.</p>
        <p>The pilot monitored 87 endpoints including 34 radio communication devices and 53 surveillance
cameras, representing approximately 5% of the simulation scale. Due to security and privacy
constraints, the pilot excluded vehicle systems and mobile devices.</p>
        <p>Real-world data quality proved substantially worse than simulation. Missing data occurred at
8.7% versus simulated 2.3%. Sensors exhibited calibration drift requiring weekly recalibration
versus simulated monthly intervals. Network interruptions were three times more frequent than
simulation, and timestamp inconsistencies occurred daily versus rarely in simulation.</p>
        <p>Despite these data quality challenges, the AI system maintained detection accuracy of 0.82
(F1score) compared to 0.89 in simulation. This degradation was expected but remained substantially
better than the baseline system's 0.71 in the same real environment, validating that the
performance advantage holds in operational conditions.</p>
        <p>Integration complexity exceeded estimates by factors of three to four. Legacy equipment lacked
modern monitoring APIs, requiring custom integration for each equipment manufacturer. Security
requirements added substantial authentication overhead, and network segregation complicated
data collection.</p>
        <p>Structured interviews with eight operators revealed mixed but ultimately positive responses.
While 73% expressed positive views about detection accuracy, only 45% initially expressed
confidence in AI recommendations. Explainability emerged as the primary concern - operators
wanted to understand why alerts occurred. After adding prototype explainability features showing
feature importance scores, operator confidence increased to 73%.</p>
        <p>This pilot deployment, while limited in scope and duration, suggests that simulation provided
reasonable performance estimates (within 10% accuracy) but significantly underestimated practical
deployment challenges around infrastructure integration and human factors requirements.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Discussion</title>
      <sec id="sec-6-1">
        <title>6.1. Performance improvements and operational impact</title>
        <p>
          The performance improvements demonstrated in evaluation translate to concrete operational
benefits. Reducing false positives from 12.3% to 3.8% means fewer unnecessary investigations and
less alert fatigue for operators. Earlier detection improves response options by enabling scheduled
maintenance during convenient times rather than emergency interventions. The attention
mechanism's interpretability provided unexpected benefits as operators learned to recognize
precursor patterns themselves [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ].
        </p>
        <p>The 72% reduction in detection time represents a significant improvement in operational
responsiveness. In law enforcement contexts where equipment failures can directly impact officer
safety, detecting problems 106 seconds faster provides valuable additional time for appropriate
response measures.</p>
      </sec>
      <sec id="sec-6-2">
        <title>6.2. Architectural considerations</title>
        <p>
          Building on established open-source tools rather than creating everything from scratch accelerated
development and leveraged extensive community testing [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. The decision to augment Prometheus
and Thanos rather than replacing them meant operators could retain familiar interfaces while
gaining AI capabilities. Separating AI components from core monitoring infrastructure created
clean boundaries simplifying development and testing.
        </p>
        <p>This architectural approach offers several advantages for law enforcement agencies. Existing
monitoring infrastructure investments remain valuable rather than requiring replacement.
Operators do not need to learn entirely new interfaces. The modular design allows agencies to
adopt AI capabilities incrementally rather than requiring complete infrastructure overhaul.</p>
      </sec>
      <sec id="sec-6-3">
        <title>6.3. Simulation methodology: strengths and limitations</title>
        <p>This research relies primarily on cloud-based simulation rather than operational law enforcement
data. This methodological choice merits explicit discussion of both strengths and limitations.
Simulation was necessary because operational law enforcement data is rarely available for research
due to security, privacy, and confidentiality constraints.</p>
        <p>
          Simulation enables controlled experimentation and reproducible research impossible with
production systems, while avoiding risks of experimental algorithms disrupting critical operational
infrastructure. The simulation was designed to maximize realism within practical constraints by
incorporating equipment failure rates matching published statistics [
          <xref ref-type="bibr" rid="ref1 ref3 ref4">1,3,4</xref>
          ] within ±5%, metric
ranges derived from manufacturer specifications, and operational patterns based on academic
literature [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] and specialist consultation.
        </p>
        <p>However, simulation inevitably simplifies reality in several important ways. Data quality in real
environments exhibits more noise, drift, and errors than simulation. Real operational contexts
contain unanticipated interactions simulation cannot model. Operator behavior and organizational
dynamics are difficult to simulate accurately. Rare edge cases may be underrepresented in
simulation.</p>
        <p>The pilot deployment results suggest that real-world performance will be 5-10% lower than
simulation results, integration effort 3-4× higher than anticipated, and maintenance requirements
elevated due to environmental variations. Nevertheless, the AI-enhanced approach still
substantially outperformed baseline methods even in real operational conditions, validating the
core value proposition despite simulation limitations.</p>
        <p>This simulation-based research represents an important first step in developing AI-enhanced
law enforcement monitoring, demonstrating feasibility and potential benefits while identifying
challenges requiring attention. Production deployment will require extensive real-world validation
with multiple agencies, robust data quality handling beyond simulation requirements, enhanced
explainability features, and comprehensive integration frameworks for legacy equipment.</p>
      </sec>
      <sec id="sec-6-4">
        <title>6.4. Explainability requirements</title>
        <p>
          Explainability remains a critical consideration. Current AI models provide limited insight into
decision-making processes, which could hinder adoption in environments requiring clear
justification for actions [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]. The pilot testing revealed that operators need not just alerts but
understandable explanations of why the system believes equipment may be failing.
        </p>
        <p>The prototype explainability features developed during this research represent initial steps
toward addressing these requirements. These features show which metrics contributed most
strongly to each alert through feature importance visualizations. However, substantial additional
work is needed to create explainability tools that operators find genuinely helpful rather than just
meeting technical requirements.</p>
      </sec>
      <sec id="sec-6-5">
        <title>6.5. Ethical considerations</title>
        <p>
          AI monitoring systems in law enforcement contexts raise ethical considerations beyond technical
effectiveness. The system's ability to track equipment usage patterns could enable undesirable
surveillance if metrics data gets misused [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. Clear policies governing appropriate use of
monitoring data and strong technical controls preventing misuse represent essential safeguards
that must accompany any deployment.
        </p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>7. Future work</title>
      <p>Validation with extensive operational data from multiple law enforcement agencies represents the
next critical step. Partnerships are being established to conduct larger-scale pilot deployments and
gather comprehensive feedback on system performance in actual operational environments. These
deployments will provide insights into practical challenges that simulation cannot capture.</p>
      <p>Explainability enhancement represents another high priority. Development of explainable AI
techniques specifically for law enforcement monitoring should explore methods that provide
transparent reasoning for anomaly detection decisions and failure predictions, enabling operators
to understand and trust system recommendations. Integration with incident management systems
and officer safety monitoring presents opportunities for expansion beyond equipment monitoring.</p>
      <p>
        Computer vision integration could enable automated quality inspection of equipment and
facilitate visual monitoring of physical infrastructure components. Federated learning approaches
offer potential for enabling knowledge sharing across multiple agencies while maintaining data
sovereignty and security [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], addressing the challenge of limited data availability in individual
organizations.
      </p>
      <p>Enhanced integration with IoT sensors and edge computing platforms could reduce latency and
enable real-time processing at distributed locations. Advanced correlation analysis between
equipment failures and operational patterns may reveal previously unknown relationships that
could further improve predictive accuracy and maintenance scheduling.</p>
    </sec>
    <sec id="sec-8">
      <title>8. Conclusions</title>
      <p>This research demonstrates that AI-enhanced monitoring systems offer significant potential for
improving law enforcement equipment reliability and operational effectiveness. The combination
of proven monitoring infrastructure (Prometheus and Thanos) with custom AI components delivers
meaningful improvements in anomaly detection, predictive maintenance, and operational
efficiency.</p>
      <p>Technical results from six-month simulation-based evaluation are encouraging, achieving
substantial improvements in detection accuracy (F1-score 0.89 vs 0.72) and response times (72%
faster) while reducing false alarms by 69%. Limited pilot deployment with one agency validated
that performance advantages hold in real operational conditions, though with expected
degradation from simulation results. The cloud-based architecture proved scalable and reliable,
suggesting sophisticated monitoring capabilities can be deployed across different agency sizes.</p>
      <p>However, important challenges require attention before widespread deployment becomes
practical. The reliance on primarily simulated data, while comprehensive and realistic, cannot fully
capture all complexities of real-world law enforcement environments. Data quality issues in
operational deployments proved worse than simulation anticipated, requiring robust preprocessing
techniques. Integration with existing systems demands careful attention to security protocols and
operational workflows. Human factors aspects need research to ensure operators can effectively
use and trust these systems. Explainability features must be enhanced to meet operational
requirements for understanding system decisions.</p>
      <p>Successful deployments will likely start with limited pilot programs focusing on specific
problems rather than attempting immediate infrastructure revolutionization. This incremental
approach allows organizations to build experience and confidence while avoiding risks associated
with large-scale changes to critical systems. The potential benefits for officer safety and operational
effectiveness make continued research important, but success requires realistic expectations and
careful attention to practical challenges of implementing AI systems in complex, security-sensitive
environments.</p>
      <p>While primary findings are based on simulated data, the combination of realistic simulation
parameters, published failure statistics, and pilot deployment validation suggests that the proposed
approaches offer promising directions for operational systems. Comprehensive real-world
validation across multiple agencies remains necessary before drawing definitive conclusions about
production effectiveness.</p>
    </sec>
    <sec id="sec-9">
      <title>Declaration on Generative AI</title>
      <p>The authors used AWS cloud services and open-source machine learning frameworks including
TensorFlow, PyTorch, and Scikit-learn to implement and validate the monitoring system. No
generative AI tools were used in writing this manuscript. All text was written by the authors who
take full responsibility for content.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <article-title>[1] International Association of Chiefs of Police, Technology Equipment Standards for Law Enforcement</article-title>
          ,
          <source>IACP Technology Standards Report</source>
          <volume>34</volume>
          (
          <year>2023</year>
          )
          <fpage>12</fpage>
          -
          <lpage>28</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          et al.,
          <article-title>Workload failure prediction for data centers</article-title>
          ,
          <source>in: IEEE Cluster Conference</source>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <article-title>[3] National Institute of Justice, Equipment Failure Analysis in Law Enforcement Operations</article-title>
          ,
          <source>NIJ Technical Report</source>
          <year>2023</year>
          -
          <volume>01</volume>
          (
          <year>2023</year>
          )
          <fpage>45</fpage>
          -
          <lpage>67</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Police</given-names>
            <surname>Executive Research Forum</surname>
          </string-name>
          ,
          <source>Technology and Law Enforcement: Current Challenges and Future Opportunities, PERF Report</source>
          (
          <year>2023</year>
          )
          <fpage>89</fpage>
          -
          <lpage>112</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>L.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          , H. Liu,
          <article-title>Quality control in defense manufacturing using machine learning</article-title>
          ,
          <source>Computers &amp; Industrial Engineering</source>
          <volume>168</volume>
          (
          <year>2022</year>
          )
          <fpage>108086</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <article-title>Predictive maintenance for security equipment: comprehensive review</article-title>
          ,
          <source>Maintenance Engineering</source>
          <volume>45</volume>
          (
          <issue>3</issue>
          ) (
          <year>2023</year>
          )
          <fpage>234</fpage>
          -
          <lpage>251</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>P.</given-names>
            <surname>Strembitskyi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yukhymchuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Lesko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Perepelytsia</surname>
          </string-name>
          ,
          <article-title>Centralized infrastructure monitoring using Thanos system</article-title>
          , Herald of Khmelnytskyi National University.
          <source>Technical Sciences 347(1)</source>
          (
          <year>2025</year>
          )
          <fpage>417</fpage>
          -
          <lpage>422</lpage>
          . https://doi.org/10.31891/
          <fpage>2307</fpage>
          -5732-2025-347-57
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M.</given-names>
            <surname>Yukhymchuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Strembitskyi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Perepelytsia</surname>
          </string-name>
          ,
          <article-title>Remote monitoring to improve production quality and efficiency</article-title>
          , Herald of Khmelnytskyi National University.
          <source>Technical Sciences</source>
          <volume>335</volume>
          (
          <issue>3- 1</issue>
          ) (
          <year>2024</year>
          )
          <fpage>330</fpage>
          -
          <lpage>334</lpage>
          . https://doi.org/10.31891/
          <fpage>2307</fpage>
          -5732-2024-335-3-44
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>V.</given-names>
            <surname>Sridharan</surname>
          </string-name>
          , Observability Engineering for Security Equipment Manufacturing,
          <string-name>
            <given-names>O</given-names>
            <surname>'Reilly Media Technical Report</surname>
          </string-name>
          (
          <year>2023</year>
          )
          <fpage>156</fpage>
          -
          <lpage>178</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>C.</given-names>
            <surname>Rodriguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Martinez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Garcia</surname>
          </string-name>
          ,
          <article-title>Reinforcement learning in security equipment manufacturing</article-title>
          ,
          <source>IEEE Transactions on Industrial Informatics</source>
          <volume>19</volume>
          (
          <issue>8</issue>
          ) (
          <year>2023</year>
          )
          <fpage>8756</fpage>
          -
          <lpage>8765</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>S.</given-names>
            <surname>García</surname>
          </string-name>
          et al.,
          <article-title>Data quality in time series: comprehensive survey</article-title>
          ,
          <source>ACM Computing Surveys</source>
          <volume>55</volume>
          (
          <issue>8</issue>
          ) (
          <year>2023</year>
          )
          <fpage>1</fpage>
          -
          <lpage>38</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>H.</given-names>
            <surname>Chen</surname>
          </string-name>
          , J. Liu,
          <article-title>Interpretable anomaly detection for critical infrastructure</article-title>
          ,
          <source>IEEE Transactions on Reliability</source>
          <volume>72</volume>
          (
          <issue>2</issue>
          ) (
          <year>2023</year>
          )
          <fpage>789</fpage>
          -
          <lpage>801</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Amazon</given-names>
            <surname>Web</surname>
          </string-name>
          <article-title>Services, Multi-account architecture best practices, AWS Architecture Center (</article-title>
          <year>2023</year>
          ). URL: https://aws.amazon.com/architecture/
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Netflix</surname>
          </string-name>
          ,
          <article-title>Chaos engineering: system resiliency in practice</article-title>
          ,
          <source>IEEE Software 40(3)</source>
          (
          <year>2023</year>
          )
          <fpage>78</fpage>
          -
          <lpage>86</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>T.</given-names>
            <surname>Chen</surname>
          </string-name>
          , C. Guestrin,
          <article-title>XGBoost: scalable tree boosting system</article-title>
          ,
          <source>ACM Transactions on Machine Learning 15(2)</source>
          (
          <year>2023</year>
          )
          <fpage>1</fpage>
          -
          <lpage>27</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>D.</given-names>
            <surname>Norman</surname>
          </string-name>
          , The Design of Everyday AI:
          <string-name>
            <surname>Human-Computer</surname>
            <given-names>Interaction</given-names>
          </string-name>
          , MIT Press,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>