<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Time series forecast of inbound call volume in call center using machine learning methods</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ruslan Krasnozhonov</string-name>
          <email>krasnozhonovr@gmail.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Aizhan Altaibek</string-name>
          <email>a.altaibek@iitu.edu.kz</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Aizhan Ydydrys</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marat Nurtas</string-name>
          <email>m.nurtas@iitu.edu.kz</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Institute of Ionosphere</institution>
          ,
          <addr-line>Gardening community IONOSPHERE 117, Almaty, 050020</addr-line>
          ,
          <country country="KZ">Kazakhstan</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>International Information Technology University</institution>
          ,
          <addr-line>34/1 Manas St., Almaty</addr-line>
          ,
          <country country="KZ">Kazakhstan</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Time series analysis involves examining data collected at various time points to identify patterns, trends, and seasonal changes. Accurate forecasting of future values is crucial for call centers to optimize staff scheduling and manage workloads effectively. While traditional statistical methods and manual forecasting have been widely used, machine learning techniques have shown promising results in enhancing forecast accuracy. This paper explores the application of machine learning for forecasting incoming call volumes, with a focus on comparing Long Short-Term Memory (LSTM) networks and Random Forest Regression models. The models are evaluated using Mean Absolute Error (MAE), Root Mean Squared Error (RMSE), and R-squared (R2) metrics. Experimental results indicate that Random Forest Regression performs well with limited data, achieving competitive MAE and RMSE values. However, LSTM networks outperform Random Forest Regression on an hourly scale, showing superior accuracy and higher R2 scores as the dataset size increases. This study demonstrates that while Random Forest Regression provides stable performance across different data sizes, LSTM models offer significant improvements in forecasting accuracy, particularly with larger datasets and high temporal granularity.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Time Series Forecasting</kwd>
        <kwd>Machine Learning</kwd>
        <kwd>LSTM</kwd>
        <kwd>Random Forest Regressor1</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Time series analysis involves examining sequences of data points collected at successive time
intervals to identify patterns, trends, and seasonal variations. This analysis is crucial for forecasting
future values, particularly in domains where understanding and predicting temporal patterns can
significantly impact decision-making and resource management. One such domain is call center
operations, where accurate forecasting of inbound call volumes is essential for optimizing staffing
levels and enhancing customer service [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>Call centers serve as a primary point of contact for customers, making effective workload
forecasting critical to prevent operational inefficiencies and improve service quality. Traditional
methods for forecasting call volumes often rely on statistical models or manual approaches, which
may not fully capture the complex and dynamic nature of call volume patterns. As call centers
continue to be a key interaction channel, there is a growing need for advanced forecasting techniques
that can handle the inherent variability and seasonality in call data.</p>
      <p>Machine learning has emerged as a powerful tool for time series forecasting, offering the ability to
model complex, non-linear relationships and capture intricate temporal dependencies. Among the
various machine learning methods, Long Short-Term Memory networks and Random Forest
Regression are particularly noteworthy for their unique strengths in time series prediction.</p>
      <p>
        LSTM networks, a type of recurrent neural network (RNN), are designed to capture long-term
dependencies and sequential patterns in data. Their ability to remember information over extended
periods makes them well-suited for modeling time series data with significant temporal dynamics [
        <xref ref-type="bibr" rid="ref2 ref3">2,
3</xref>
        ]. LSTMs have been shown to outperform traditional statistical methods in capturing complex
patterns and trends, especially in cases with strong seasonality and non-linear relationships [
        <xref ref-type="bibr" rid="ref4 ref5 ref6">4, 5, 6</xref>
        ].
      </p>
      <p>
        On the other hand, Random Forest Regression is an ensemble learning technique that aggregates
multiple decision trees to improve prediction accuracy and robustness. Random Forest Regression
excels at handling both linear and non-linear relationships and is known for its ability to manage
high-dimensional data and prevent overfitting [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8</xref>
        ]. Its flexibility and ease of implementation make
it a popular choice for various forecasting tasks, including those involving time series data [
        <xref ref-type="bibr" rid="ref10 ref11 ref12 ref9">9-12</xref>
        ].
      </p>
      <p>
        This study aims to explore and compare the performance of Long Short-Term Memory and
Random Forest Regression in forecasting inbound call volumes for call centers. The comparison is
driven by the need to evaluate how each model's strengths contribute to accurate predictions in this
specific context. By assessing their performance based on Mean Absolute Error (MAE), Root Mean
Squared Error (RMSE), and R-squared (R2) metrics, we seek to provide actionable insights into the
most effective forecasting methods for call center operations [
        <xref ref-type="bibr" rid="ref13 ref14">13-16</xref>
        ].
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Problem statement</title>
      <p>In the dynamic environment of call center operations, accurately predicting inbound call volumes is
essential for efficient resource management and maintaining customer satisfaction. Traditional
forecasting methods often struggle to account for the inherent variability and complexity in call
volume data, leading to challenges in optimizing agent scheduling and controlling operational costs.</p>
      <p>Despite the potential of machine learning models like Random Forest Regression and Long
ShortTerm Memory networks to address these challenges, their application in the specific context of call
center call volume forecasting remains insufficiently explored. There is a critical need to determine
the effectiveness of these models in capturing the intricate temporal patterns in call volume data,
particularly when forecasting across different time scales, such as daily and half-hour intervals.</p>
      <p>This study seeks to address the following key research questions:


</p>
      <p>How do Random Forest Regression and LSTM models compare in terms of accuracy and
reliability when forecasting inbound call volumes in a call center?
What are the specific conditions under which each model performs optimally?
How do different time scales (daily vs. half-hour intervals) and feature engineering strategies
impact the performance of these models?</p>
      <p>By answering these questions, this research aims to provide a deeper understanding of how
machine learning can be effectively leveraged to improve call center operations through more precise
call volume forecasting.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>This study employs two machine learning models—Random Forest Regression and Long Short-Term
Memory networks—to forecast inbound call volumes in a call center. The methods are applied to two
datasets: one aggregated by day, capturing total daily call volumes, and the other segmented by
30minute intervals, providing a finer granularity of call activity throughout the day. The
methodological approach is divided into several key stages.</p>
      <sec id="sec-3-1">
        <title>3.1. Data preprocessing and feature engineering</title>
        <p>The dataset used in this study spans from April 2022 to August 2024 and includes detailed records of
inbound call volumes. The datasets were first examined for missing values and outliers. Any
anomalies were either corrected or removed to ensure data quality. An anomaly detection step was
performed using the Isolation Forest algorithm. This method is particularly effective in identifying
outliers in high-dimensional data. The model was trained on the scaled data with a contamination
rate set to 0.001, meaning that 0.1% of the data points were considered as anomalies. These identified
anomalies were then removed from the dataset, resulting in a cleaner dataset for subsequent
modeling.</p>
        <p>The plot illustrates the identified anomalies (marked as red points) in the dataset, which were
detected using the Isolation Forest algorithm with a contamination rate of 0.001.</p>
        <p>To enhance the predictive performance of the machine learning models, we employed a variety of
feature engineering techniques based on the available timestamped call data. The raw data provided
timestamps for each inbound call, which we transformed into meaningful features aimed at capturing
both temporal patterns and the operational dynamics of the call center.</p>
        <p>First, we generated time-based features by extracting key temporal information from the
timestamp data. Each call record was assigned its corresponding half-hour of the day (e.g., 9:30, 10:00)
and the day of the week (e.g., Monday, Tuesday), thereby allowing the model to account for daily and
weekly patterns. We also introduced binary indicators to signal specific periods of interest. For
instance, a binary variable was created to mark whether a call occurred during daytime half-hours
(8:00 AM to 9:30 PM), which typically see higher call volumes. Similarly, we defined variables for
lunch hours (12:00 PM to 1:00 PM) and work hours (8:00 AM to 6:00 PM) to capture fluctuations
related to these timeframes.</p>
        <p>Categorical features, such as the day of the week, were one-hot encoded to ensure that the model
did not mistakenly treat these categories as having an inherent order. This encoding technique
created separate binary variables for each day, allowing the model to capture variations in call
volume across different days without assuming any linear relationship between them.</p>
        <p>In addition to these temporal features, we incorporated lag features to account for the influence of
past call volumes on future predictions. These lag features represent call volumes at previous time
steps, and we generated lags from one to ten time steps prior. This allowed the model to recognize
autocorrelations within the data, such as whether a high call volume in the previous period would
lead to a similar or opposite trend in the next period.</p>
        <p>To further capture patterns and trends in the call volume data, we calculated rolling statistics.
Rolling mean and rolling standard deviation were computed over various time windows to smooth
out short-term fluctuations and reveal longer-term trends. For instance, the rolling mean over the
past 8 steps provided the model with information about the general trend in call volume, while the
rolling standard over the past 4 step deviation highlighted the variability within those windows.</p>
        <p>By thoroughly transforming the timestamp data into informative features, we aimed to improve
the model's ability to predict future call volumes by leveraging both short-term dependencies and
longer-term trends.</p>
        <sec id="sec-3-1-1">
          <title>Feature Type</title>
        </sec>
        <sec id="sec-3-1-2">
          <title>Lag Features</title>
        </sec>
        <sec id="sec-3-1-3">
          <title>Dataset Type</title>
        </sec>
        <sec id="sec-3-1-4">
          <title>Daily and Half-Hourly</title>
        </sec>
        <sec id="sec-3-1-5">
          <title>Rolling Statistics</title>
        </sec>
        <sec id="sec-3-1-6">
          <title>Daily and Half-Hourly</title>
        </sec>
        <sec id="sec-3-1-7">
          <title>Categorical Variables</title>
        </sec>
        <sec id="sec-3-1-8">
          <title>Daily and Half-Hourly</title>
        </sec>
        <sec id="sec-3-1-9">
          <title>Description</title>
        </sec>
        <sec id="sec-3-1-10">
          <title>Previous call volumes at lagged time steps</title>
        </sec>
        <sec id="sec-3-1-11">
          <title>Measures of variability and trend over rolling windows</title>
        </sec>
        <sec id="sec-3-1-12">
          <title>Features representing categorical aspects such as time of day, day of week, etc.</title>
          <p>3.2.</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>Model training and evaluation</title>
        <p>Random Forest Regression model was chosen for its robustness and ability to capture nonlinear
relationships in the data. The model was trained using the preprocessed feature set. Hyperparameters
such as the number of trees, maximum depth, and minimum samples per split were optimized using
RandomizedSearchCV. These hyperparameters were selected to balance model complexity and
performance, ensuring that the model generalizes well to unseen data.</p>
        <p>LSTM networks were chosen for their strength in modeling sequential data and capturing
longterm dependencies, which are crucial for accurate time series forecasting. The LSTM model includes
multiple LSTM layers to capture temporal dependencies, followed by dense layers to map the output
to the target variable. The model architecture is as follows:


</p>
        <p>Input layer with shape matching the feature dimensions;
The model architecture includes multiple LSTM layers, with a dropout layer to prevent
overfitting;</p>
        <p>Dense output layer with a single unit for regression.</p>
        <p>Key hyperparameters such as the number of LSTM units, learning rate, and batch size were tuned.
Early stopping was used to monitor validation loss and prevent overfitting.</p>
        <p>Both models were evaluated using cross-validation to ensure generalizability. The data was split
into training and validation sets, with the models trained on the training set and evaluated on the
validation set.</p>
        <p>The models were assessed using metrics such as Mean Absolute Error (MAE) and Root Mean
Squared Error (RMSE) to measure their predictive accuracy and explainability.</p>
        <p>A comparative analysis was conducted to evaluate the performance of the Random Forest and
LSTM models across different time scales (daily vs. 30-minute intervals) and feature sets. This
analysis aimed to identify the conditions under which each model performs best.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Implementation and tools</title>
        <p>The models were implemented using Python, with libraries such as Scikit-learn for the Random
Forest model and TensorFlow/Keras for the LSTM model. Data scaling and anomaly detection were
performed using StandardScaler and IsolationForest, respectively.</p>
        <p>The training and evaluation processes were conducted in a high-performance computing
environment to handle the computational demands of model training, particularly for the LSTM
network.</p>
        <p>Through this methodological approach, the study seeks to provide a robust comparison between
traditional ensemble methods and deep learning techniques for forecasting inbound call volumes,
offering insights into the most effective strategies for call center management.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results and discussion</title>
      <sec id="sec-4-1">
        <title>4.1. Model Performance</title>
        <p>The Random Forest model demonstrated strong performance across both daily and half-hourly
datasets. Using the optimized hyperparameters, the model achieved an average Mean Absolute Error
(MAE) of 481 for the daily dataset and 14 for the half-hourly dataset. The Root Mean Squared Error
(RMSE) was 674 for daily and 26 for half-hourly data. The R-squared (R²) values were 0.8958 for daily
and 0.9686 for half-hourly data, indicating a good fit for both datasets.</p>
        <p>The LSTM model, leveraging its capacity to capture long-term dependencies in sequential data,
also performed well. The average MAE was 387 for the daily dataset and 13 for the half-hourly
dataset. The RMSE was 595 for daily and 23 for half-hourly data. The R-squared (R²) values were
0.9350 for daily and 0.9720 for half-hourly data. The LSTM model showed superior performance in
capturing temporal patterns and trends compared to the Random Forest model.</p>
        <p>The LSTM model's training dynamics, visualized through loss graphs, indicated efficient learning
and convergence, highlighting the model's ability to adapt to sequential data over epochs. This
reinforces the LSTM model's suitability for time series forecasting.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Comparison across time scales</title>
        <p>The comparative analysis revealed that both models performed better with the half-hourly dataset
compared to the daily dataset. This improvement is likely due to the finer granularity of the
halfhourly data, which allows the models to capture more detailed patterns and trends in call volumes.
The LSTM model, in particular, showed a significant advantage in handling the higher resolution of
the half-hourly data, reflecting its strength in modeling sequential dependencies.</p>
        <p>Figure 3 illustrates the predicted vs. actual call volumes for both models, demonstrating the
models' effectiveness in capturing call volume trends.</p>
        <p>This study highlights the strengths and limitations of Random Forest Regression and LSTM
networks in forecasting inbound call volumes. The results indicate that both models offer valuable
insights, with LSTM networks showing particular strength in handling sequential data and capturing
long-term dependencies. The findings suggest that a combination of these models, along with
continuous refinement of features and parameters, can provide robust forecasting solutions for call
center operations.</p>
        <p>The residuals plot displays the difference between actual and predicted values against the
predicted values for both Random Forest Regression and Long Short-Term Memory models. It shows
a funnel-shaped pattern where residuals widen with higher predicted values, indicating that Random
Forest Regression tends to have larger errors for higher call volumes, while LSTM maintains tighter
residuals and more consistent predictions. This suggests that LSTM may be more reliable for
forecasting larger call volumes compared to Random Forest Regression. Both models perform
similarly well for smaller predicted values.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Feature importance</title>
        <p>The feature importance analysis for the Random Forest model indicated that lag features (e.g., Lag-1,
Lag-2) and rolling statistics (e.g., Standard Deviation, Rolling Mean) were the most influential in
predicting call volumes. This aligns with the expectation that recent call volumes and historical
trends play a crucial role in forecasting.</p>
        <p>For the LSTM model, the impact of features was less straightforward due to the model's ability to
learn complex temporal patterns. However, lagged features and rolling statistics were still important,
as they provided essential context for the LSTM's sequential processing.
4.4.</p>
      </sec>
      <sec id="sec-4-4">
        <title>Model limitations</title>
        <p>While the Random Forest model performed well, it is limited by its inability to capture very long-term
dependencies due to its non-sequential nature. Additionally, the model's performance can degrade if
the feature set does not adequately capture all relevant temporal patterns.</p>
        <p>The LSTM model, despite its strengths, requires significant computational resources and can be
sensitive to hyperparameter settings. The performance of the LSTM model also depends on the
quality and granularity of the input features.</p>
      </sec>
      <sec id="sec-4-5">
        <title>4.5. Practical implications</title>
        <p>The insights from this study have been applied in a call center setting, where machine learning
models are now used to forecast call volumes and optimize staffing levels. While the initial
implementation has shown promise in improving staffing efficiency and reducing operational costs,
we are still in the process of fully evaluating the impact on service levels and customer satisfaction.
Early results suggest that these models are aligning staffing more closely with expected call activity,
potentially leading to reduced wait times and more effective resource management.</p>
        <p>Future enhancements could involve integrating additional factors into the forecasting models,
such as external events, promotional campaigns, and technical anomalies. Addressing these aspects
could further refine the accuracy of the predictions and improve overall model performance.
Exploring other advanced machine learning techniques or hybrid models may also offer further
improvements in forecasting capabilities, providing even more robust solutions for call center
management and beyond.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>This study evaluated the effectiveness of Random Forest Regression and Long Short-Term Memory
models for forecasting inbound call volumes in call centers. The analysis was conducted using both
daily and half-hourly datasets to determine the models' performance across different time scales.</p>
      <p>The LSTM model outperformed the Random Forest model in forecasting call volumes. Specifically,
the LSTM achieved lower Mean Absolute Error (MAE) and Root Mean Squared Error (RMSE), and
higher R-squared (R2) values, demonstrating its superior ability to capture long-term dependencies
and sequential patterns in time series data.</p>
      <p>The training loss graph for the LSTM model illustrated efficient learning and convergence,
reinforcing its suitability for time series forecasting.</p>
      <p>While both models showed strong performance, the LSTM's ability to model temporal trends and
dependencies provided a significant advantage over the Random Forest model, especially for
halfhourly data.</p>
      <p>Accurate call volume forecasting is essential for effective call center management. The results
suggest that LSTM models can enhance staffing decisions and operational efficiency by providing
more precise predictions compared to traditional models like Random Forest Regression.</p>
      <p>Further research could explore additional machine learning techniques and integrate external
factors to improve forecasting accuracy. Expanding the scope to include varied time series
characteristics may also offer deeper insights into call volume prediction.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <sec id="sec-6-1">
        <title>The authors have not employed any Generative AI tools.</title>
        <p>International Conference on Engineering &amp; MIS 2020
doi: https://doi.org/10.1145/3410352.3410778.
[15] Nurtas M., Baishemirov Zh., Zhanabekov Zh. (2020). Convalutional Neural Networks as a
method to solve estimation problem of acoustic wave propogation in poroelastic media, News of
the National Academy of sciences of the Republic of Kazakhstan. 4(332).</p>
        <p>doi:  https://doi.org/10.32014/2020.2518-1726.65.
[16] Nurtas M., Baishemirov Zh., Zhanabekov Zh. (2020). Applying Neural Network for predicting
cardiovascular disase risk, News of the National Academy of sciences of the Republic of
Kazakhstan.4(332). Doi: https://doi.org/10.32014/2020.2518-1726.62.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Ibrahim</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ye</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>L</surname>
          </string-name>
          'Ecuyer,
          <string-name>
            <given-names>P.</given-names>
            , &amp;
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <surname>H.</surname>
          </string-name>
          (
          <year>2016</year>
          ).
          <article-title>Modeling and forecasting call center arrivals: A literature survey and a case study</article-title>
          .
          <source>International Journal of Forecasting</source>
          ,
          <volume>32</volume>
          (
          <issue>3</issue>
          ),
          <fpage>865</fpage>
          -
          <lpage>874</lpage>
          . doi :
          <volume>10</volume>
          .1016/j.ijforecast.
          <year>2015</year>
          .
          <volume>11</volume>
          .012.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Hochreiter</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Schmidhuber</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          (
          <year>1997</year>
          ).
          <article-title>Long Short-Term Memory</article-title>
          .
          <source>Neural Computation</source>
          ,
          <volume>9</volume>
          (
          <issue>8</issue>
          ),
          <fpage>1735</fpage>
          -
          <lpage>1780</lpage>
          . doi:
          <volume>10</volume>
          .1162/neco.
          <year>1997</year>
          .
          <volume>9</volume>
          .8.1735.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Gers</surname>
            ,
            <given-names>F. A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schmidhuber</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Cummins</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          (
          <year>2000</year>
          ).
          <article-title>Learning to forget: continual prediction with LSTM</article-title>
          .
          <source>Neural Computation</source>
          ,
          <volume>12</volume>
          (
          <issue>10</issue>
          ),
          <fpage>2451</fpage>
          -
          <lpage>2471</lpage>
          . doi:
          <volume>10</volume>
          .1162/089976600300015015.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Kumar</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tripathi</surname>
            ,
            <given-names>B. K.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Singh</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          (
          <year>2023</year>
          ).
          <article-title>Attention-based LSTM network-assisted time series forecasting models for petroleum production</article-title>
          .
          <source>Engineering Applications of Artificial Intelligence</source>
          ,
          <volume>123</volume>
          , 106440. doi:
          <volume>10</volume>
          .1016/j.engappai.
          <year>2023</year>
          .
          <volume>106440</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Graves</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Schmidhuber</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          (
          <year>2005</year>
          ).
          <article-title>Framewise Phoneme Classification with Bidirectional LSTM Networks</article-title>
          .
          <source>IEEE Transactions on Neural Networks</source>
          ,
          <volume>14</volume>
          (
          <issue>5</issue>
          ),
          <fpage>993</fpage>
          -
          <lpage>998</lpage>
          . doi:
          <volume>10</volume>
          .1109/IJCNN.
          <year>2005</year>
          .
          <volume>1556215</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Lindemann</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Müller</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vietz</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jazdi</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Weyrich</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          (
          <year>2021</year>
          ).
          <article-title>A survey on long short-term memory networks for time series prediction</article-title>
          .
          <source>Procedia CIRP</source>
          ,
          <volume>99</volume>
          ,
          <fpage>650</fpage>
          -
          <lpage>655</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.procir.
          <year>2021</year>
          .
          <volume>03</volume>
          .088.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Breiman</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          (
          <year>2001</year>
          ).
          <source>Random Forests. Machine Learning</source>
          ,
          <volume>45</volume>
          (
          <issue>1</issue>
          ),
          <fpage>5</fpage>
          -
          <lpage>32</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Liaw</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Wiener</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          (
          <year>2002</year>
          ).
          <article-title>Classification and Regression by RandomForest</article-title>
          .
          <source>R News</source>
          ,
          <volume>2</volume>
          (
          <issue>3</issue>
          ),
          <fpage>18</fpage>
          -
          <lpage>22</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Qi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          (
          <year>2005</year>
          ).
          <article-title>Neural Network Forecasting of Stock Market Time Series</article-title>
          .
          <source>European Journal of Operational Research</source>
          ,
          <volume>160</volume>
          (
          <issue>2</issue>
          ),
          <fpage>570</fpage>
          -
          <lpage>584</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.ejor.
          <year>2003</year>
          .
          <volume>08</volume>
          .037.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Lachaud</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Adam</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,, &amp;
          <string-name>
            <surname>Miskovic</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          (
          <year>2023</year>
          ).
          <article-title>Comparative Study of Random Forest and Support Vector Machine Algorithms in Mineral Prospectivity Mapping with Limited Training Data</article-title>
          . Minerals,
          <volume>13</volume>
          (
          <issue>8</issue>
          ),
          <fpage>1073</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Nurtas</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhantaev</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Altaibek</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          (
          <year>2024</year>
          ),
          <article-title>Earthquake time-series forecast in Kazakhstan territory: Forecasting accuracy with SARIMAX</article-title>
          .
          <source>Procedia Computer Science</source>
          ,
          <volume>231</volume>
          ,
          <fpage>353</fpage>
          -
          <lpage>358</lpage>
          .doi:
          <volume>10</volume>
          .1016/j.procs.
          <year>2023</year>
          .
          <volume>12</volume>
          .216.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Aizhan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tokhtakhunov</surname>
          </string-name>
          , Il',.. &amp;
          <string-name>
            <surname>Nurtas</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          (
          <year>2024</year>
          ).
          <article-title>The Efficacy of Autoencoders in the Utilization of Tabular Data for Classification Tasks</article-title>
          .
          <source>Procedia Computer Science</source>
          ,
          <volume>238</volume>
          . doi:
          <volume>10</volume>
          .1016/j.procs.
          <year>2024</year>
          .
          <volume>06</volume>
          .052.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Nurtas</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhantaev</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Altaibek</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nurakynov</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mekebayev</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shiyapov</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Iskakov</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ydyrys</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          (
          <year>2023</year>
          ).
          <article-title>Predicting the Likelihood of an Earthquake by Leveraging Volumetric Statistical Data through Machine Learning Techniques</article-title>
          .
          <source>Engineered Science</source>
          ,
          <volume>26</volume>
          (
          <issue>1031</issue>
          ). doi:
          <volume>10</volume>
          .30919/es1031.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Nurtas</surname>
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ydyrys</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Altaibek</surname>
            <given-names>A.</given-names>
          </string-name>
          (
          <year>2020</year>
          ).
          <article-title> Using of Machine Learning algorithm and Spectral method for simulation of Nonlinear Wave Equation</article-title>
          .
          <source> ICEMIS'20: Proceedings of the 6th</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>