<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>adaptive cache management⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Victoria Vysotska</string-name>
          <email>victoria.a.vysotska@lpnu.ua</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kyrylo Smelyakov</string-name>
          <email>kyrylo.smelyakov@nure.ua</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Anastasiya Chupryna</string-name>
          <email>anastasiya.chupryna@nure.ua</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Matvii Kuchapin</string-name>
          <email>matvii.kuchapin@nure.ua</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Kharkiv National University of Internal Affairs</institution>
          ,
          <addr-line>L. Landau Avenue 27 61080 Kharkiv</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Kharkiv National University of Radio Electronics</institution>
          ,
          <addr-line>Nauky Ave. 14, Kharkiv, 61166</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>Caching effectiveness in practice depends less on which policy is chosen and more on how its parameters are configured. Static settings rarely track non-stationary workloads, while existing adaptive TTLs and hybrid eviction schemes are under-adopted because there is no reproducible way to select parameters. Building on prior work that introduced adaptive TTL and multi-criteria eviction, this paper shifts the focus from algorithm design to reproducible configuration. We formulate parameter selection as a multicriteria decision process that maps workload characteristics and operational constraints to scenario-based presets and deterministic correction rules. The methodology provides engine-agnostic guidance for latency-sensitive APIs, reuse-oriented analytics, edge/IoT deployments, and mixed web services, including recommended ranges for lifetime bounds and eviction weights. Rather than reporting new experiments, the contribution consolidates prior empirical insights into practitioner-oriented, workload-aware presets and a lightweight operational validation checklist. The aim is to reduce trial-and-error, improve reproducibility across cache engines (e.g., Redis, Memcached), and lay groundwork for automated, selftuning caching systems that maintain performance as workloads evolve.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;adaptive cache management</kwd>
        <kwd>parameter selection</kwd>
        <kwd>time-to-live (TTL)</kwd>
        <kwd>eviction policies</kwd>
        <kwd>multi-criteria decision making</kwd>
        <kwd>workload-aware caching</kwd>
        <kwd>performance optimization</kwd>
        <kwd>reproducibility 1</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Caching mechanisms are widely employed in data-intensive and high-load systems to reduce
latency and pressure on primary storage. However, the effectiveness of such mechanisms depends
largely not only on the chosen policy, such as TTL, LRU, LFU, but also on the configuration of key
parameters that control these policies. Static parameterization often cannot adapt to workload
dynamics. Overly conservative lifetimes lead to inefficient memory usage, while aggressive eviction
reduces hit rates and increases tail latency. The presence of heterogeneous data objects,
nonstationary access distributions, and strict resource constraints further complicates parameter
selection.</p>
      <p>
        Adaptive and hybrid caching strategies have been proposed to address these challenges [
        <xref ref-type="bibr" rid="ref1 ref2 ref3">1-3</xref>
        ].
Nevertheless, existing research primarily focuses on the design of algorithms rather than on
systematic approaches to parameter tuning. Unlike algorithmic innovation, which introduces new
eviction or TTL strategies, the contribution of this work lies in defining how such mechanisms can
be systematically configured. In other words, the focus shifts from what algorithms to use to how
they should be parameterized in order to match diverse workloads and constraints.
      </p>
      <p>
        This paper addresses the identified gap. Building on our earlier work [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], which introduced an
adaptive cache management strategy that integrates exponential-decay frequency, recency-aware
aging, and a multi-criteria eviction score under different workload conditions, the present paper
emphasizes configuration rather than new algorithm design. That prior study reported gains in hit
ratio and eviction stability compared to LRU and LFU, but also revealed that cache performance
remained highly sensitive to parameter choices such as TTL bounds and weight distributions. This
sensitivity provides the motivation for the current study: to develop a reproducible methodology
for selecting cache parameters in accordance with workload characteristics and operational
constraints, thereby reducing reliance on repeated trial-and-error.
      </p>
      <p>The proposed methodology formulates parameter selection as a multi-criteria decision-making
process. It considers how workload properties such as access asymmetry, volatility, read/write
ratio, and object size distribution influence the relative importance of parameters including
recency, frequency, and memory size. It also incorporates operational constraints such as
servicelevel requirements, memory capacity, and processor load, and shows how scenario-based
recommendations can be used to generalize prior findings into practical guidelines
The main contributions of this paper are as follows:



a formalized framework for parameterization of adaptive cache management, independent
of a specific cache engine and compatible with exponential-decay TTL models and
multicriteria eviction scoring;
scenario-based guidelines, including tabular recommendations, for selecting weighting
coefficients (wa , wf , wm) and lifetime boundaries (T base , T max) in different deployment
contexts;
a lightweight protocol for deployment and tuning, incorporating profiling, initial
calibration, and runtime monitoring, designed to minimize repeated manual adjustments.</p>
      <p>In summary, this study advances adaptive cache management from the level of algorithmic
design to that of a methodological framework for parameter selection. By codifying parameter
choice as a structured process rather than an ad hoc practice, the work facilitates faster adoption in
practical systems and establishes a foundation for future research on automated, self-tuning cache
management. Beyond its theoretical contribution, the methodology aims to reduce operational
trial-and-error and provide practitioners with engine-agnostic, scenario-driven rules that accelerate
time-to-value. This practical focus ensures that adaptive caching strategies can be deployed more
reliably in production without requiring exhaustive tuning for each new workload.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Previous studies have revealed both the potential and limitations of adaptive caching. In our earlier
study [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], we presented an adaptive strategy that combines dynamic TTL assignment with
exponential decay and a multi-criteria eviction score that balances freshness, frequency, and
memory consumption. Evaluation demonstrated improvements in hit rate, memory utilization, and
eviction stability compared to LRU, LFU, and random eviction, but also revealed a critical
limitation: cache performance was highly sensitive to parameter tuning, particularly weighting
factors and TTL boundaries. This observation motivated the present work, which shifts the focus
from algorithmic design to the development of a systematic methodology for parameter selection
for heterogeneous scenarios.
      </p>
      <p>
        Classic replacement policies such as LRU and LFU are still widely used due to their simplicity
and low overhead. However, their reliance on a single criterion often leads to suboptimal behavior
under dynamic loads. TTL-based approaches provide additional control in cases where data validity
depends on time, but static assignments often result in premature deletion or prolonged storage of
stale items. Adaptive TTL variants, such as exponential decay functions and workload-aware
policies, increase flexibility but still lack generalized parameter selection rules and often depend on
special tuning [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>
        Hybrid strategies attempt to combine several criteria to increase adaptability. Examples include
composite relevance and frequency scores [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] or adaptive replacement caches (ARC) that respond
to changes in popularity distribution [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Despite their effectiveness, these approaches primarily
emphasize algorithmic novelty rather than systematic parameterization.
      </p>
      <p>
        More recent work continues to refine adaptive eviction in specialized contexts. CAKE
introduces layer-aware eviction indicators tailored for large language models [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], while MadaKV
adapts eviction weights for multimodal inference [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Both highlight the role of adaptive weighting
but remain domain-specific. Similarly, SIEVE (2024) provides a lightweight algorithm that improves
miss ratios relative to LRU [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], and Cold-RL (2025) applies reinforcement learning to cache
eviction in NGINX by incorporating multiple features while retaining LRU fallback for stability
[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
      </p>
      <p>
        Survey studies, such as Advancements in Cache Management: A Review of Machine Learning
Techniques for Cache Replacement [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], highlight the growing research interest in adaptive and
learning-based caching. However, this review also identifies a persistent shortcoming: the lack of
reproducible, scenario-driven methodologies for parameter configuration, which continues to limit
practical adoption. Similar reproducibility challenges have also been observed in applied ML
domains such as road accident detection [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
      </p>
      <p>In summary, while the literature demonstrates significant advances in adaptive and hybrid
cache management, it consistently leaves unresolved the problem of parameter selection. The
present work addresses this gap by introducing a methodological framework that formalizes
parameter configuration as a reproducible process, thereby extending prior algorithmic
contributions with systematic and practice-oriented guidance.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodological Framework</title>
      <sec id="sec-3-1">
        <title>3.1. Problem Statement</title>
        <p>
          The effectiveness of adaptive cache management strategies strongly depends on how their
parameters are configured. While prior research has introduced algorithms that combine recency,
frequency, and memory consumption into unified eviction scores, as well as adaptive TTL
mechanisms with exponential decay functions [
          <xref ref-type="bibr" rid="ref5 ref6 ref7">5-7</xref>
          ], the selection of configuration parameters
remains largely unsystematic. In particular, two groups of parameters are critical:


weighting coefficients (wa , wf , wm), which determine the relative importance of access
recency, frequency of use, and memory footprint within the eviction score;
lifetime boundaries (T base , T max), which control the dynamic adjustment of TTL values and
thereby influence data freshness and resource utilization.
        </p>
        <p>Improper configuration of these parameters can lead to significant performance degradation.
For example, assigning excessive weight to recency may cause frequent eviction of long-term
popular items, while overemphasizing frequency may result in stale objects occupying memory for
extended periods. Likewise, overly narrow TTL ranges can trigger high eviction rates and write
amplification, whereas excessively broad ranges may cause cache pollution by retaining low-value
entries.</p>
        <p>The challenge, therefore, lies in formulating parameter selection as a multi-criteria
decisionmaking problem, in which system designers must balance several conflicting objectives. These
include cache hit ratio, read and write latency, memory utilization, and CPU overhead.
Furthermore, the importance of each objective varies depending on deployment scenarios: web
applications require balanced responsiveness, real-time systems prioritize data freshness, analytics
pipelines emphasize frequency-based reuse, and IoT or embedded devices operate under stringent
memory constraints.</p>
        <p>
          Despite substantial progress in adaptive caching research [
          <xref ref-type="bibr" rid="ref1 ref12 ref2 ref3">1-3, 12</xref>
          ], the literature still lacks a
reproducible methodology for parameter selection that generalizes across heterogeneous
workloads. Current approaches typically rely on empirical trial-and-error, workload-specific
heuristics, or machine-learning models that require extensive retraining. This absence of
systematic guidelines hinders both reproducibility and portability of adaptive cache mechanisms.
        </p>
        <p>The present work addresses this gap by developing a methodological framework for parameter
selection in adaptive cache management. The framework provides scenario-driven rules, supported
by empirical evidence, enabling practitioners to configure (wa , wf , wm) and (T base , T max) effectively
without repeated costly experimentation. In doing so, it establishes one of the first reproducible
methodologies that bridges empirical evaluation with scenario-based parameterization, advancing
adaptive caching from ad hoc tuning toward systematic deployment.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Adaptive Lifetime Model</title>
        <p>
          The adaptive lifetime model determines the residency duration of cache entries as a function of
recent access activity. Static TTL assignments are prone to inefficiencies, while the adaptive
approach provides workload-sensitive adjustment. In our earlier study [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ], the model was
formalized as
        </p>
        <p>
          TTL=T base+(T max−T base)× f n ,
(1)
where TTL is the adaptive time-to-live assigned to a cache entry, T base is the guaranteed
minimum residency time (lower bound), T max is the maximum residency time (upper bound), and
f n∈ [
          <xref ref-type="bibr" rid="ref1">0,1</xref>
          ] is the normalized access activity score, typically computed using exponential decay with
λ ≥ 0.
        </p>
        <p>The parameters T base and T max provide explicit control over the lower and upper bounds of
residency, while the method of computing f n through exponential decay introduces responsiveness
to workload volatility. A shorter λ for the decay function makes the system adapt quickly to recent
changes, whereas a longer λ produces smoother, more stable behavior.</p>
        <p>The adaptive lifetime model is considered in this work as a fundamental component whose
effectiveness depends primarily on how its parameters are configured, rather than on the novelty
of its formulation.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Adaptive Eviction Model</title>
        <p>
          When cache capacity is reached, the system must decide which entries to discard. Single-criterion
rules such as LRU or LFU are often insufficient under heterogeneous or time-varying workloads. To
capture multiple objectives within one decision, a unified eviction score was introduced in earlier
work [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] and is recalled here as a tunable component rather than re-derived in detail:
S=wa × 1/(an+ ε )+ wf × f n+ wm × mmax / m , wa ​, wf ​, wm ​≥ 0 , wa ​+ wf ​+ wm ​=1 ,
(2)
where S is the eviction score of an entry (higher implies lower priority for retention);
wa , wf , wm are non-negative weights for recency, frequency, and memory footprint (typically
normalized); an is the normalized access-age term; f n is the normalized frequency term; m is the
entry’s memory size (bytes); mmax is the maximum entry size observed (bytes); ε is a small constant
preventing division by zero.
        </p>
        <p>By adjusting the relative weights, the model can approximate established policies such as LRU
when recency is emphasized, LFU when frequency is prioritized, or size-aware strategies when
memory footprint dominates. At the same time, it supports intermediate hybrids that balance these
objectives. This flexibility provides continuity with classical approaches while extending them into
a broader design space where trade-offs can be explicitly managed. Within the scope of this study,
the eviction score is considered not as a new algorithm but as a configurable mechanism whose
effectiveness depends on systematic parameterization aligned with workload characteristics and
resource constraints. In combination with the adaptive lifetime model, it forms one of the two core
building blocks of the proposed methodological framework, supporting reproducible and
scenariodriven cache management.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Parameterization Approach</title>
        <p>The parameterization of adaptive cache models requires a systematic and reproducible procedure
that translates workload characteristics into effective configurations.The approach is presented as
an operational protocol rather than as a new experiment.</p>
        <p>
          The process begins with scenario identification, where workloads are profiled using available
telemetry data and domain expertise. Characteristics such as freshness versus reuse orientation,
capacity load, access pattern variability, read–write asymmetry, and object size distribution are
analyzed to assign the deployment to a representative scenario class (e.g., latency-sensitive API,
burst-loaded service, or memory-constrained device). This classification determines the subsequent
parameter choices. Comparable methodology for scenario-based parameterization has been
demonstrated in comparative studies of clustering algorithms for market segmentation [
          <xref ref-type="bibr" rid="ref14 ref15 ref16 ref17 ref18">14-18</xref>
          ].
        </p>
        <p>Based on the identified scenario, initial parameters are selected for the adaptive lifetime model,
the effective freshness sensitivity determined by the λ of the activity signal, and the eviction
weights. These seed values are drawn from scenario-based presets and adjusted against operational
constraints such as global lower bounds on minimum lifetime, namespace-specific upper bounds,
and projected memory utilization.</p>
        <p>To verify plausibility, the configuration undergoes a lightweight validation checklist rather than
controlled experiments. This checklist typically includes:



ensuring that latency percentiles remain within service-level targets;
confirming that eviction and write rates do not exceed acceptable thresholds;
checking that CPU and memory overhead remain within budget.</p>
        <p>When deviations are observed, deterministic rules are applied. Instead of iterative fine-tuning,
structured mappings between observed metric deviations and parameter adjustments provide
stability and prevent oscillations. Since lifetime and eviction mechanisms interact, adjustments are
coordinated to preserve trade-offs between reuse, responsiveness, and memory efficiency.</p>
        <p>Once parameters are refined, the configuration is documented in a scenario-to-parameter log,
recording the scenario label, final parameter values, justification for adjustments, and observed
operational metrics. This log builds a cumulative knowledge base that supports reproducible
configuration of future deployments without repeating manual trial-and-error.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Operational Validation Checklist</title>
      <p>The use of predefined scenarios in adaptive cache management requires a structured mechanism to
confirm that the selected parameters behave as expected in real operating conditions. Since no new
experiments are proposed in this paper, the purpose of this section is not to describe a controlled
evaluation protocol in detail, but to provide a checklist for operational validation. This checklist is
intended to be used as a reproducible and simple process that allows system developers and
practitioners to verify that a given configuration meets service level requirements, resource
constraints, and stability criteria in production environments. By shifting the focus from
experimental measurements to operational validation, the methodology ensures reliable and
consistent deployment of adaptive cache management across different systems.</p>
      <p>The first element of operational validation concerns compliance with service level objectives.
Cache configuration directly affects response times, especially at higher percentiles such as P95 and
P99, which are often critical in latency-sensitive environments. When applying preset settings
based on scenarios, operators must confirm that these latency thresholds remain within the
acceptable values defined by the application domain. Hit rate stability is equally important, as it
reflects the cache's ability to deliver consistent benefits across varying traffic intensities. Unlike
testing, which measures improvements relative to a baseline, operational validation only requires
assurance that system-level goals will not be compromised after the configuration is deployed.</p>
      <p>The second aspect of validation relates to the efficient use of resources. Caching is often subject
to strict limitations on memory and processor power, especially in multi-user or embedded
environments. For this reason, operational monitoring must confirm that load levels remain
balanced and that memory is neither underutilized nor overloaded. Excessive eviction frequency,
which manifests as uncontrolled outflow, can lead to increased write amplification and ultimately
reduce the expected performance benefits of caching. Similarly, if eviction policies are not
configured properly, large objects can monopolize cache capacity and displace smaller, frequently
used records. Therefore, validation includes verifying that memory allocation is proportional, that
the eviction rate remains below acceptable thresholds, and that the CPU load caused by caching
operations does not interfere with other critical tasks on the same system.</p>
      <p>The third element of the checklist involves corrective adjustment using deterministic rules.
Since the methodology defines parameterization as a multi-criteria decision-making process, each
parameter is linked to observable system metrics that indicate when adjustment is required. For
example, if eviction rates exceed acceptable limits, increasing the lower bound of the lifetime (T base
) stabilizes cache residency. If stale objects are observed, decreasing the upper bound (T max)
restores freshness. If large objects constantly occupy most of the capacity, increasing the weight
assigned to memory volume (wm) in the displacement estimate prevents monopolization. These
adjustments are not discovered through iterative trial and error, but are prescribed by structured
comparisons between observed deviations and parameter updates. Thus, the validation process
avoids fluctuations and maintains system stability.</p>
      <p>Documentation is an important component of operational validation. Every setting and every
decision must be recorded in a log of scenarios and parameters, which reflects the context of the
workload, the settings selected, the deviations observed, the corrective actions taken, and the
metrics obtained. Such documentation ensures that the knowledge gained during one deployment
can be transferred to subsequent deployments without having to go through the same verification
process again. Over time, the accumulation of these records forms a knowledge base that enhances
reproducibility and allows for faster adaptation to changes in workloads. It is important to note
that this approach is engine-independent and can be applied to various caching platforms, such as
Redis, Memcached, or in-memory caching layers built into microservice architectures.</p>
      <p>Thus, the operational validation checklist transforms parameter tuning from an informal and
potentially unstable practice into a structured process based on service level compliance, resource
efficiency, corrective adjustments, and documentation. Unlike experimental evaluation, which is
designed to confirm performance improvements over baseline metrics, the checklist is designed to
ensure the safety of practical implementation by ensuring that scenario settings are consistent with
actual workloads and constraints. This emphasis on reproducibility and portability reinforces the
broader methodological contribution of the paper and provides practitioners with a pragmatic path
for implementing adaptive cache management in production systems.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Scenario-Based Guidelines</title>
      <p>The operational validation checklist outlined in Section 4 ensures that cache configurations remain
compliant with service-level objectives and resource constraints once deployed. However,
practitioners also require clear entry points for selecting parameters before validation can take
place. To address this, the following section translates the methodological framework into
reproducible scenario-based guidelines that specify recommended ranges for lifetime boundaries
and eviction weights. These guidelines do not claim to provide globally optimal values; rather, they
function as practical presets that reduce reliance on ad hoc tuning and accelerate adoption in
production systems.</p>
      <p>The rationale for scenario-based recommendations arises from the diversity of application
domains in which caching is deployed. Latency-sensitive APIs, analytics pipelines, edge or IoT
environments, and heterogeneous microservice platforms impose distinct requirements that cannot
be satisfied by a single configuration. By mapping these domains to representative scenarios and
prescribing parameter presets tailored to their dominant characteristics, the methodology bridges
theoretical models with operational practice. The subsections that follow present such mappings
together with explanatory notes, allowing practitioners to select the most relevant profile and then
confirm its suitability using the operational checklist.</p>
      <sec id="sec-5-1">
        <title>5.1. Latency-Sensitive APIs</title>
        <p>Latency-sensitive APIs and interactive microservices are characterized by bursty traffic and strict
Service-Level Objectives (SLOs) concerning tail latency. In these environments, cache configuration
directly affects response times, especially at higher percentiles such as P95 and P99.Consequently,
eviction decisions must prioritize recency (w_a) to maintain responsiveness, while lifetime bounds
(T base, T max) are kept narrow to ensure rapid churn and prevent memory saturation.</p>
        <p>Typical configurations set the guaranteed minimum residency time (T base) between 30 and 60
seconds, which ensures minimum residency but still allows rapid churn. The maximum residency
time (T max) is typically set between 300 and 600 seconds, limiting the maximum lifetime to avoid
cache saturation. The decay coefficient λ is set relatively high, between 0.3 and 0.5, because a
higher λ accelerates aging and makes the system adapt quickly to recent changes, thereby
prioritizing recent access activity. The eviction weights (wa , wf , wm) generally emphasize recency,
with values around 0.6–0.7, 0.2–0.3, 0.1–0.2, respectively, to meet latency SLOs. Recommended
parameter ranges are summarized in Table 1.</p>
        <p>This recency-first configuration is essential for meeting strict tail latency SLOs, as it ensures the
cache adapts rapidly to bursty traffic and prioritizes the most recently accessed items. While this
aggressive eviction of older items might slightly lower the overall hit ratio by removing potentially
reusable (but less recent) entries, this trade-off is necessary to maintain the responsiveness required
in latency-sensitive applications.</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Reuse-Oriented Analytics</title>
        <p>Analytical workloads, including recommendation engines, are characterized by intensive reuse of
historical data. In such scenarios, frequency plays a dominant role, and longer retention periods are
required to ensure the availability of valuable elements. The objective of this configuration is to
maximize the cache hit ratio and minimize recomputation overhead.</p>
        <p>The guaranteed minimum residency time (T base) is set in the range of 5 to 10 minutes,
specifically to avoid premature eviction of items that are useful in the long term. The maximum
residency time (T max) is set high, ranging from 1 to 4 hours, enabling long-term reuse. The decay
coefficient λ is set low, in the range of 0.05–0.1, because a lower λ produces smoother, more stable
behavior that preserves long-term popularity trends and smooths out fluctuations. Eviction weights
(wa , wf , wm) are shifted strongly towards frequency, with values around 0.2–0.3, 0.6–0.7, 0.1–0.2.
Recommended parameter ranges are shown in Table 2.</p>
        <p>This configuration maximizes the hit ratio and reduces overhead for recalculation, although it
may increase the risk of cache pollution if the access distribution changes unexpectedly. When
latency requirements become critical, the recency weight can be moderately increased to restore
responsiveness.</p>
        <p>
          This parameterization reflects the focus of analytical workloads on reuse and is consistent with
findings from prior evaluation [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] and with caching strategies described in industry and academic
studies.
        </p>
      </sec>
      <sec id="sec-5-3">
        <title>5.3. Edge and IoT Environments</title>
        <p>Edge and IoT workloads are typically subject to stringent memory constraints and exhibit high
variability in access patterns. Caching in these domains is often subject to strict limitations on
memory and processor power. To prevent a small number of large objects from monopolizing the
limited cache capacity, eviction decisions must place strong emphasis on the memory footprint.</p>
        <p>Typical T base values are set low, 10–30 seconds, and T max is set between 120–300 seconds. These
short bounds are used to ensure rapid rotation, keeping memory available for rotation and
preventing long-lived objects from persisting. The decay coefficient λ is set higher between 0.4–0.6,
enforcing rapid aging, which is suitable for volatile access patterns. Eviction weights ( wa , wf , wm)
are skewed toward memory, with distributions around 0.2–0.3, 0.2–0.3, 0.5–0.6. Recommended
configurations are summarized in Table 3. Such settings prevent oversized entries from saturating
limited capacity, although they may reduce hit ratios due to more frequent evictions. This trade-off
is acceptable in contexts where predictability and fairness take precedence over maximizing reuse.</p>
        <p>
          These guidelines reflect practical strategies reported in prior evaluation [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] and in lightweight
caching approaches for embedded and IoT systems. The focus is on predictable performance under
capacity limitations, even at the expense of reduced hit ratios.
        </p>
      </sec>
      <sec id="sec-5-4">
        <title>5.4. Mixed Web and Microservices</title>
        <p>Heterogeneous environments, such as large-scale web platforms and multi-tenant microservice
architectures, require balanced performance across multiple dimensions, including responsiveness,
reuse, and memory efficiency. These workloads demand a setting where no single factor dominates
and resilience is prioritized across scenarios.</p>
        <p>A representative configuration uses moderate parameter values. T base is set between 1–2
minutes, which prevents rapid churn without excessive retention. T max is set higher, between 20–
40 minutes, providing stability for moderately reused items. The decay coefficient λ is set to an
intermediate range of 0.1–0.3. This intermediate λ is selected specifically to balance responsiveness
(recency) and stability (frequency). Eviction weights (wa , wf , wm) are distributed relatively evenly,
with 0.3–0.4, 0.4–0.5, and 0.2–0.3, respectively. Recommended configurations are summarized in
Table 4.</p>
        <p>This balanced setup provides robustness under diverse and shifting workloads, without
aggressively prioritizing a single criterion. Continuous monitoring and periodic reassessment
remain essential to ensure sustained alignment with workload dynamics.</p>
        <p>
          These ranges are presented as pragmatic compromises between responsiveness and reuse. They
are informed by prior evaluation [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] and supported by reports on multi-tenant web and
microservice workloads, where no single factor dominates and resilience across scenarios is
prioritized.
        </p>
        <sec id="sec-5-4-1">
          <title>6. Results and Discussion</title>
          <p>
            The proposed methodology consolidates insights from prior evaluation [
            <xref ref-type="bibr" rid="ref4">4</xref>
            ] and published reports
into a structured framework for parameter selection in adaptive caching. Rather than presenting
new measurements, the emphasis lies in explaining how scenario-based presets are expected to
behave under representative workload conditions. This section therefore interprets the rationale
and anticipated effects of applying the guidelines introduced in Section 5, showing how systematic
parameterization can shape cache behavior across different domains.
          </p>
          <p>For latency-sensitive APIs, configurations dominated by recency weights are expected to
improve responsiveness and reduce tail latency, even though they may shorten the residency of
items that are useful in the long term. In reuse-oriented analytics, frequency-oriented
parameterizations are likely to promote higher hit ratios and minimize recomputation overhead,
though they may adapt more slowly to sudden shifts in access distributions. Edge and IoT
workloads illustrate how memory-aware settings are anticipated to provide fairness and
predictability by preventing large objects from monopolizing capacity, even if this results in lower
aggregate hit ratios. Finally, in mixed web and microservice environments, balanced parameter
ranges are expected to enhance robustness, ensuring that no single objective dominates and that
performance remains stable under heterogeneous and shifting conditions. The deterministic
adjustment rules embedded in the methodology provide an additional stabilizing effect. Instead of
iterative fine-tuning, operators can rely on structured mappings between observed deviations such
as excessive eviction rates or latency violations and parameter updates. This rule-based approach is
expected to reduce oscillations and facilitate convergence toward suitable configurations.</p>
          <p>Overall, the discussion supports the central premise that adaptive cache management benefits
not only from algorithmic design but also from systematic parameterization. By coupling
scenariodriven presets with lightweight operational validation, the methodology is anticipated to reduce
trial-and-error, improve reproducibility, and enhance transferability across cache engines. In doing
so, it offers a pathway for more reliable deployment of adaptive caching strategies in practice.</p>
        </sec>
        <sec id="sec-5-4-2">
          <title>7. Conclusion</title>
          <p>
            This paper has advanced adaptive cache management by shifting the focus from algorithm design
to systematic parameterization. Building on prior evaluations [
            <xref ref-type="bibr" rid="ref4">4</xref>
            ] and insights from existing
literature, it introduced a methodological framework that treats parameter selection as a
reproducible process. The framework integrates two configurable components an adaptive lifetime
model and a multi-criteria eviction model into scenario-driven guidelines and deterministic
correction rules.
          </p>
          <p>The guidelines provide practitioners with actionable starting points for configuring cache
systems under diverse conditions, including latency-sensitive APIs, reuse-oriented analytics,
memory-constrained IoT environments, and heterogeneous web or microservice platforms. Rather
than prescribing globally optimal configurations, the recommendations function as reproducible
presets that reduce reliance on ad hoc trial-and-error and accelerate practical adoption.</p>
          <p>The broader contribution of the methodology lies in highlighting reproducibility,
interpretability, and transferability as essential elements of cache management, alongside raw
performance. By documenting parameter choices as scenario-to-parameter mappings and by
embedding lightweight operational validation, the approach enables knowledge accumulation
across deployments and supports engine-agnostic use in systems such as Redis or Memcached.</p>
          <p>Future work may extend this foundation by exploring automated scenario detection, machine
learning–based refinement of parameter mappings, and continuous self-tuning mechanisms that
adapt to workload evolution. In this way, adaptive caching can progress toward becoming a robust,
self-managing subsystem of modern data-intensive architectures.</p>
        </sec>
        <sec id="sec-5-4-3">
          <title>Declaration on Generative AI</title>
          <p>The authors don’t employed any Generative AI tools.</p>
        </sec>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Srivatsa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Fasfous</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. Anh</given-names>
            <surname>Vu Doan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Wild</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Herkersdorf</surname>
          </string-name>
          ,
          <article-title>Exploring a Hybrid Votingbased Eviction Policy for Caches and Sparse Directories on Manycore Architectures</article-title>
          , Volume
          <volume>87</volume>
          , Microprocessors and Microsystems,
          <year>2021</year>
          . doi:
          <volume>10</volume>
          .1016/j.micpro.
          <year>2021</year>
          .
          <volume>104384</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Shi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Fan</surname>
          </string-name>
          , Optimized Caching Strategy:
          <article-title>A Hybrid of Least Recently Used and Least Frequently Used Methods, 5th International Conference on Computer Network Security and Software Engineering (CNSSE '25), Association for Computing Machinery</article-title>
          , New York, NY, USA,
          <year>2025</year>
          , pp.
          <fpage>133</fpage>
          -
          <lpage>140</lpage>
          . doi:
          <volume>10</volume>
          .1145/3732365.3732389.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Sharif</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. H. Y.</given-names>
            <surname>Moghaddam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. A. H.</given-names>
            <surname>Seno</surname>
          </string-name>
          ,
          <article-title>A hybrid Bi-level management framework for caching and communication in Edge-AI enabled IoT</article-title>
          ,
          <source>Journal of Network and Computer Applications</source>
          , Volume
          <volume>232</volume>
          ,
          <year>2024</year>
          . doi:
          <volume>10</volume>
          .1016/j.jnca.
          <year>2024</year>
          .
          <volume>104000</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>O.</given-names>
            <surname>Shevchenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kuchapin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Dudar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Shirokopetleva</surname>
          </string-name>
          ,
          <source>Enhancing Redis Cache Efficiency Based on Dynamic TTL and Adaptive Eviction Mechanism</source>
          ,
          <source>2025 IEEE Open Conference of Electrical, Electronic and Information Sciences (eStream)</source>
          , Vilnius, Lithuania,
          <year>2025</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          , doi: 10.1109/eStream66938.
          <year>2025</year>
          .
          <volume>11016870</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Basu</surname>
          </string-name>
          , et al,
          <article-title>Adaptive TTL-Based Caching for Content Delivery</article-title>
          , IEEE/ACM Transactions on Networking, Volume
          <volume>26</volume>
          , pp.
          <fpage>1063</fpage>
          -
          <lpage>1077</lpage>
          ,
          <year>2018</year>
          , doi: 10.1109/TNET.
          <year>2018</year>
          .
          <volume>2818468</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Srivatsa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Nagel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Fasfous</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. Anh</given-names>
            <surname>Vu Doan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Wild</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Herkersdorf,
          <article-title>HyVE: A Hybrid Voting-based Eviction Policy for Caches</article-title>
          , 
          <source>2020 IEEE Nordic Circuits and Systems Conference (NorCAS)</source>
          , Oslo, Norway,
          <year>2020</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>7</lpage>
          , doi: 10.1109/NorCAS51424.
          <year>2020</year>
          .
          <volume>9265136</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>P.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kannaujia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Sarma</surname>
          </string-name>
          ,
          <source>Adaptive Replacement Cache Policy in Named Data Networking</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>5</lpage>
          . doi:
          <volume>10</volume>
          .1109/CONIT51480.
          <year>2021</year>
          .
          <volume>9498489</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Qin</surname>
          </string-name>
          , et al,
          <article-title>CAKE: Cascading and Adaptive KV Cache Eviction with Layer Preferences</article-title>
          ,
          <string-name>
            <surname>ICLR</surname>
          </string-name>
          <year>2025</year>
          , arXiv,
          <year>2025</year>
          . doi:
          <volume>10</volume>
          .48550/arXiv.2503.12491.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>K.</given-names>
            <surname>Li</surname>
          </string-name>
          , et al,
          <article-title>MadaKV: Adaptive Modality-Perception KV Cache Eviction for Efficient Multimodal Long-Context Inference</article-title>
          , arXiv,
          <year>2025</year>
          . doi:
          <volume>10</volume>
          .48550/arXiv.2506.15724.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , et al,
          <article-title>SIEVE is simpler than LRU: an efficient turn-key eviction algorithm for web caches</article-title>
          ,
          <source>21st USENIX Symposium on Networked Systems Design and Implementation (NSDI'24)</source>
          , USENIX Association, USA,
          <year>2024</year>
          , pp.
          <fpage>1229</fpage>
          -
          <lpage>1246</lpage>
          . doi:
          <volume>10</volume>
          .5555/3691825.3691893.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Aa. Gupta</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Bhayani</surname>
          </string-name>
          , Cold-RL:
          <article-title>Learning Cache Eviction with Offline Reinforcement Learning for NGINX</article-title>
          , arXiv,
          <year>2025</year>
          . doi:
          <volume>10</volume>
          .48550/arXiv.2508.12485.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>K.</given-names>
            <surname>Keshav</surname>
          </string-name>
          ,
          <article-title>Advancements in cache management: a review of machine learning innovations for enhanced performance and security</article-title>
          ,
          <source>Frontiers in Artificial Intelligence</source>
          , Volume
          <volume>8</volume>
          ,
          <year>2025</year>
          . doi:
          <volume>10</volume>
          .3389/frai.
          <year>2025</year>
          .
          <volume>1441250</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>O.</given-names>
            <surname>Byzkrovnyi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Savulioniene</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Smelyakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Sakalys</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Chupryna</surname>
          </string-name>
          , Comparison of Potential Road Accident Detection Approaches, Vide Tehnologija Resursi - Environment, Technology Resources, Volume
          <volume>3</volume>
          , pp.
          <fpage>50</fpage>
          -
          <lpage>55</lpage>
          . doi:
          <volume>10</volume>
          .17770/etr2023vol3.
          <fpage>7299</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>D.</given-names>
            <surname>Teslenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sorokina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Smelyakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Filipov</surname>
          </string-name>
          ,
          <article-title>Comparative Analysis of the Applicability of Five Clustering Algorithms for Market Segmentation</article-title>
          ,
          <source>2023 IEEE Open Conference of Electrical, Electronic and Information Sciences (eStream)</source>
          , Vilnius, Lithuania,
          <year>2023</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          , doi: 10.1109/eStream59056.
          <year>2023</year>
          .
          <volume>10134796</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>V.</given-names>
            <surname>Vysotska</surname>
          </string-name>
          ,
          <article-title>Computer linguistic system modelling for Ukrainian language processing</article-title>
          ,
          <source>in: CEUR Workshop Proceedings</source>
          , vol.
          <volume>3722</volume>
          ,
          <year>2024</year>
          , pp.
          <fpage>288</fpage>
          -
          <lpage>342</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>V.</given-names>
            <surname>Vysotska</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Kyrychenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Demchuk</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gruzdo</surname>
          </string-name>
          ,
          <article-title>Holistic adaptive optimization techniques for distributed data streaming systems</article-title>
          ,
          <source>in: CEUR Workshop Proceedings</source>
          , vol.
          <volume>3668</volume>
          ,
          <year>2024</year>
          , pp.
          <fpage>120</fpage>
          -
          <lpage>132</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>V.</given-names>
            <surname>Lytvyn</surname>
          </string-name>
          , et al,
          <article-title>Development of intellectual system for data de-duplication and distribution in cloud storage</article-title>
          ,
          <source>Webology</source>
          <volume>16</volume>
          (
          <issue>2</issue>
          ) (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>V.</given-names>
            <surname>Vysotska</surname>
          </string-name>
          , et al,
          <article-title>Information resource management technology based on fuzzy logic</article-title>
          ,
          <source>in: Intellectual Systems of Decision Making and Problems of Computational Intelligence</source>
          , Springer, Cham,
          <year>2020</year>
          , pp.
          <fpage>164</fpage>
          -
          <lpage>182</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>