<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Information Control Systems &amp; Technologies, September</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Validation of a Neural Network Architecture for Approximating an Analytical Model of Eye Condition</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Vladimir Vychuzhanin</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nickolay Rudnichenko</string-name>
          <email>nickolay.rud@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alexey Vychuzhanin</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Olga Guzun</string-name>
          <email>olga.v.guzun@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Odesa</institution>
          ,
          <addr-line>65001</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Odessa Polytechnic National University</institution>
          ,
          <addr-line>Shevchenko Avenue 1, Odessa, 65001</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>2</volume>
      <fpage>4</fpage>
      <lpage>26</lpage>
      <abstract>
        <p>This paper presents an intelligent deep learning-based model for comprehensive analysis of the human eye's condition. The developed neural network system integrates key ophthalmological parameters, including intraocular pressure, volumetric blood circulation, visual acuity, visual field index, and perfusion pressure, along with additional factors such as age, vascular health, and genetic predisposition. A synthetic dataset of 250,000 samples was generated based on clinically observed parameter ranges from the Filatov Institute of Eye Diseases and Tissue Therapy. This controlled dataset enabled architectural validation of a neural network model designed to approximate a physiologically meaningful function (Seye). Although real patient data were not used, the study demonstrates the feasibility of building a robust diagnostic framework, laying the groundwork for future application to clinical datasets. The neural network architecture includes three hidden layers with ReLU activation, ensuring high prediction accuracy. Model evaluation demonstrated a high coefficient of determination and low values of root mean squared error and mean absolute percentage error, indicating a strong correlation between predicted and actual values. The obtained results confirm the potential of neural network methods for automated eye condition analysis. The proposed model can be applied for early diagnosis and monitoring of ophthalmological diseases, as well as a decision-support tool in clinical practice. Future work includes integrating real medical data to enhance the model's generalizability and developing hybrid approaches that combine traditional mathematical methods with deep learning.</p>
      </abstract>
      <kwd-group>
        <kwd>machine learning</kwd>
        <kwd>neural network</kwd>
        <kwd>ophthalmology</kwd>
        <kwd>prediction</kwd>
        <kwd>eye condition modeling 1</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Ophthalmological research plays a vital role in preserving vision, as the condition of the eye is
influenced by a wide range of interacting physiological and clinical parameters, including
intraocular pressure, volumetric blood circulation, visual acuity, visual field index, tear production,
and perfusion pressure [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ]. Traditional diagnostic methods often rely on simplified models such
as regression equations or differential systems. While useful, these approaches may not fully
capture the complex, nonlinear relationships between variables and often depend on subjective
interpretation [
        <xref ref-type="bibr" rid="ref1 ref2 ref5">1, 2, 5</xref>
        ]. This limits their ability to integrate multiple heterogeneous factors and
reduces predictive accuracy in practical settings. In recent years, machine learning and deep
learning methods have gained considerable attention in ophthalmology, providing powerful tools
for analyzing large volumes of data and identifying hidden patterns in visual and clinical indicators
[
        <xref ref-type="bibr" rid="ref3 ref4 ref5 ref6">3, 4, 5, 6</xref>
        ]. These models enable more objective and data-driven decision-making. However, most
existing studies are focused on the detection of specific diseases or classification tasks, without
network-based approach to approximating a physiologically informed analytical model of overall
eye condition (Seye), constructed from expert knowledge and clinical reasoning [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8, 9</xref>
        ]. The
function Seye integrates key ophthalmological parameters using nonlinear transformations that
reflect known physiological dependencies. Importantly, the current stage of research focuses on
validating the neural network architecture under idealized, controlled conditions using a
synthetically generated dataset. The target variable is calculated analytically to ensure that the
model structure and training process are reliable and stable. This methodological step lays the
groundwork for future application to real clinical data, where variability, noise, and missing values
will present greater challenges [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8, 9</xref>
        ].
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Literature review and existing methods</title>
      <p>2.1.</p>
      <sec id="sec-2-1">
        <title>Traditional methods for eye condition modeling</title>
        <p>
          In the early stages of ophthalmic modeling, traditional methods such as regression analysis and
differential equations were widely used. These approaches enabled the description of key processes
such as intraocular pressure fluctuations, blood flow dynamics, and the relationship between
physiological parameters and visual function [
          <xref ref-type="bibr" rid="ref9">10, 11</xref>
          ]. For instance, in [10], mathematical models are
presented that describe the dynamics of intraocular pressure using fluid balance equations, allowing
for the simulation of steady-state and oscillatory pressure regimes under various external
influences. In [
          <xref ref-type="bibr" rid="ref9">11</xref>
          ], the importance of choosing the correct unit of analysis (e.g., one or both eyes) is
emphasized, as it directly impacts the validity of statistical inference. However, these traditional
approaches face several significant limitations: subjectivity in interpretation, due to reliance on
manual analysis and expert opinion [
          <xref ref-type="bibr" rid="ref10">12</xref>
          ]; limited parameter coverage, as most models incorporate
only a few variables and ignore other potentially important clinical or biometric indicators [
          <xref ref-type="bibr" rid="ref9">11</xref>
          ];
insufficient capacity to capture nonlinear dependencies, which is crucial given the multifactorial
and dynamic nature of ocular processes. These limitations have fueled a transition toward more
powerful and flexible data-driven methods.
2.2.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Modern methods and neural network architectures</title>
        <p>
          CNNs are widely used in ophthalmology for analyzing retinal fundus images, OCT scans, and
fundus photography [
          <xref ref-type="bibr" rid="ref11 ref12 ref13">13, 14, 15</xref>
          ]. These models can automatically extract clinically significant
features such as microaneurysms, neovascularization, and retinal detachment, often without
requiring manual feature engineering. Studies such as [
          <xref ref-type="bibr" rid="ref11 ref12">13, 14</xref>
          ] have demonstrated CNN-based
systems achieving diagnostic performance comparable to human experts. In [
          <xref ref-type="bibr" rid="ref11">13</xref>
          ], transfer learning
is used to improve model accuracy, especially when training data is limited. RNNs, including LSTM
and GRU variants, are applied to analyze temporal dynamics in ophthalmological measurements
such as intraocular pressure trends, blood flow oscillations, and visual field progression [
          <xref ref-type="bibr" rid="ref14 ref15">16, 17</xref>
          ]. For
example, the RNN model in [
          <xref ref-type="bibr" rid="ref14">16</xref>
          ] enables real-time prediction of eye condition changes, making it
useful for longitudinal disease monitoring (e.g., glaucoma or age-related macular degeneration).
Such approaches are particularly valuable for detecting early pathological changes across multiple
examination points.
2.3.
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>Hybrid and ensemble models</title>
        <p>
          Many modern studies implement hybrid models that combine classical statistical techniques with
machine learning algorithms. These models offer improved robustness, adaptability, and
performance in the presence of noisy, incomplete, or multimodal data [
          <xref ref-type="bibr" rid="ref16 ref17 ref18">18, 19, 20</xref>
          ]. In [
          <xref ref-type="bibr" rid="ref16">18</xref>
          ], a
structure combining SVMs and CNNs is described, while [
          <xref ref-type="bibr" rid="ref18">20</xref>
          ] introduces ensemble methods
combining gradient boosting and neural networks to reduce overfitting and improve
generalizability.
        </p>
        <p>
          Study [
          <xref ref-type="bibr" rid="ref17">19</xref>
          ] also highlights the integration of therapeutic and biometric data for more holistic eye
condition modeling.
        </p>
      </sec>
      <sec id="sec-2-4">
        <title>Interpretability and generalization</title>
        <p>
          Despite the complexity of neural architectures, increasing attention is being paid to interpretability,
which is crucial in clinical applications. Works such as [23, 24] focus on creating interpretable AI
systems capable of explaining the contribution of each parameter to the final prediction.
Furthermore, [
          <xref ref-type="bibr" rid="ref20">22</xref>
          ] demonstrates the potential of hybrid modeling-merging simulation-based and
machine learning-based approaches which is especially effective when dealing with limited data or
highly complex physiological systems.
2.5.
        </p>
      </sec>
      <sec id="sec-2-5">
        <title>Comparative analysis of methods</title>
        <p>
          Modern neural network methods demonstrate high accuracy in classification and prediction tasks in
ophthalmology, significantly outperforming classical algorithms in identifying complex patterns
within high-dimensional data. For example, Moradi et al. [
          <xref ref-type="bibr" rid="ref18">20</xref>
          ] showed that deep ensemble learning
achieves superior performance in the automated classification of early-stage age-related macular
degeneration (AMD), offering better generalization than individual models. Meanwhile, Archana et
al. [
          <xref ref-type="bibr" rid="ref19">21</xref>
          ] reviewed the application of traditional machine learning algorithms for glaucoma detection,
highlighting that, despite their simplicity and interpretability, classical models lag behind deep
learning methods when processing nonlinear and heterogeneous ophthalmic data. A prominent
trend in current research is the development of hybrid models and stacked generalization
techniques, which combine the strengths of various deep architectures. Kaushik et al. [25], for
instance, proposed a stacked ensemble of convolutional neural networks for diabetic retinopathy
diagnosis, achieving both high accuracy and robustness across diverse datasets.
        </p>
        <p>Similarly, Kansal et al. [26] introduced a visual feature embedding and selection method that
significantly enhances model performance for ocular disease classification. Vidivelli et al. [27]
focused on strategies for optimizing deep learning architectures tailored to ophthalmological
diagnostics.</p>
        <p>Their work emphasizes domain-specific attention mechanisms, advanced regularization
techniques, and model calibration to improve interpretability and reliability in clinical
decisionmaking. In summary, contemporary research underscores the promise of hybrid and explainable AI
systems that combine the strong feature extraction capabilities of deep learning with the
transparency and robustness of classical approaches. This integration is pivotal for the advancement
of intelligent decision support systems in ophthalmology and for increasing trust among clinical
practitioners.
2.6.</p>
      </sec>
      <sec id="sec-2-6">
        <title>Research objectives and tasks</title>
        <p>Objective: develop a neural network model for accurately predicting the overall condition of the
eye. Tasks:
• Collect and preprocess data (clinical), including normalization and outlier removal;
• Develop a mathematical model that reflects nonlinear dependencies;
• Design an optimal neural network architecture (potentially incorporating CNN and RNN
elements) and ensemble models;
• Train and validate the model using RMSE, MAPE, and R² metrics;
• Analyze the contribution of individual parameters to enhance model interpretability;.
• its advantages and
future potential.</p>
        <p>Modern neural network methods offer automatic feature extraction, high prediction accuracy,
and the ability to integrate heterogeneous data. However, challenges such as the need for large
datasets and limited interpretability remain. Addressing these issues is key to developing a
universal model capable of effectively assessing eye health.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Research methodology</title>
      <p>3.1.
Key ophthalmological parameters were selected for developing the eye condition model [28]: IOP
(Intraocular Pressure) a primary indicator of ocular tone and a crucial diagnostic criterion for
glaucoma and other eye diseases; RQ (Volumetric Intraocular Circulation) reflects ocular blood
supply, with impairments indicating potential vascular pathologies; BCVA (Best Corrected Visual
Acuity) a standard measure of visual function used for diagnosing and monitoring treatment
effectiveness; Tr (Tear Production)
diagnosing dry eye syndrome; VFI (Visual Field Index) assesses the extent of visual field
preservation, particularly relevant for monitoring glaucoma progression; Pperf (Perfusion Pressure)
reflects the efficiency of blood flow in ocular tissues, with impairments potentially leading to
vision disorders; Additional Parameters (Age, Vascular Condition, Genetic Predisposition)
considered as significant risk factors for ophthalmological diseases. The interdependencies of the
parameters characterizing eye condition are presented in Table 1.</p>
      <sec id="sec-3-1">
        <title>Loss of visual fields (VFI) is linked to progressive decline in visual acuity (BCVA) (e.g.. glaucoma, high myopia).</title>
      </sec>
      <sec id="sec-3-2">
        <title>Aging leads to reduced intraocular blood flow and narrowing of visual fields due to vascular degeneration.</title>
      </sec>
      <sec id="sec-3-3">
        <title>Tear production declines with age. leading to dry eye syndrome, negatively affecting quality of life.</title>
      </sec>
      <sec id="sec-3-4">
        <title>Vascular tone changes affect retinal blood flow, contributing to degenerative retinal diseases (e.g.. glaucoma, diabetic retinopathy).</title>
      </sec>
      <sec id="sec-3-5">
        <title>Alters overall eye condition and affects patients' quality of life.</title>
        <p>To develop the eye condition model, a set of key ophthalmological parameters proposed in [28]
was used, including IOP, RQ, BCVA, Tr, VFI, and Pperf. These parameters are recognized as
l state. The combination of
physiological and individual factors provides a comprehensive representation of eye health,
enabling the construction of an accurate predictive model. The primary variables used in the model,
along with their normal values and variation ranges, are as follows: IOP - normal: 10 21 mmHg,
Range - 5 60 mmHg; RQ - normal: 3.2 3.5%; Range - 0.5 9.0%; BCVA - normal: 1.0, Range - 0 2.0;
Tr - normal: 10 30 mm, Range: 1 40 mm; VFI - normal: 100%; Pperf - normal: 55 80 mmHg; Range
20 100 mmHg%; additional parameters: age, vascular condition, genetic factors, etc.
3.2.
logarithmic, exponential, and quadratic terms) reflects the known physiological relationships
between ophthalmic parameters and functional eye condition. For instance, intraocular pressure
and perfusion pressure are known to affect ocular health nonlinearly, which justifies the use of
logarithmic terms. Coefficients k1 through k9, as well as modifying terms A, B, and functions such
as henv, hstress, i(lifestyle), and j(genetic factors) were introduced to account for interaction effects
and individual variability. Their form was selected based on preliminary simulations, clinical
interpretability, and their effectiveness in model calibration. While these coefficients were initially
set heuristically, they were subsequently optimized using the training dataset during the neural
network fitting process, allowing the network to approximate the nonlinear dependencies encoded
in the original equation. The formula for calculating the target variable Seye is presented,
incorporating nonlinear dependencies (logarithmic, exponential, and quadratic) on each parameter.
The formulas are structured into separate blocks for better readability. The Seye model describes
eye health as a function of key ophthalmological parameters, ensuring accurate diagnosis and
prediction</p>
        <p>Seye = k1 ⋅ log(IOP + 1) + A + k2 ⋅ log(RQ + 1) + B + k3 ⋅ (BCVA − foffcet)2 ∙ A + k4 ∙
∙ log(Tr + 1) ∙ henv(environmentlafactor) + k5 ⋅ e−VF1 ∙ A + k6 ∙ log(1 + Pperf) ∙ B + k7 ∙
⋅ log(1 +
) + hstress(stresslevel, activitylevel) + k8 ⋅ log(age + 1) ∙ i(lifestyle) + k9 ∙


∙  (


where  =  (
α
t1</p>
        <p>= 
f(age, 
g(
henv( 
 
, 


, 

, 



);
 ℎ</p>
        <p>⋅  
 ℎ
ℎ
),
)</p>
        <p>(1)
 ℎ
ℎ
)=1+0.1⋅age−0.05⋅</p>
        <p>ℎ)=1+0.2⋅log(</p>
        <p>)=1+0.3⋅exp(−0.01⋅ 
+1)−0.1⋅ 


hstress( stress_level,activity_level)=1+0.05⋅stress_level−0.02⋅activity_level
interactions and clinical observations. This semi-empirical approach forms the foundation for model
training. Interpretation of the Model Equation. he model equation for Seye represents a weighted
sum of logarithmic, exponential, and quadratic transformations of key ophthalmological and
physiological parameters. It incorporates: core clinical measurements (e.g., IOP, RQ, BCVA, Tr, VFI,
Pperf); modifier functions based on age, vascular health, stress, environment, and lifestyle;
interacting terms that reflect how multiple health factors together affect the overall eye condition.
This structure allows the model to capture complex nonlinear interactions between biological
systems and external influences in a mathematically tractable form (Figure 1).</p>
        <sec id="sec-3-5-1">
          <title>Neural network model development</title>
          <p>To model eye conditions based on key ophthalmological parameters, a clinical dataset was created.
The process includes the following steps:</p>
          <p>1. Defining input parameter ranges real clinical ranges were established for each key
parameter, including intraocular pressure (IOP), volumetric intraocular circulation (RQ), best
corrected visual acuity (BCVA), tear production (Tr), visual field index (VFI), perfusion pressure
(Pperf), as well as additional factors such as age, vascular condition, and other variables. For
example, IOP is in the range of 5 60 mmHg, RQ from 0.5 to 9.0, BCVA from 0 to 2.0, etc;
2. Generating clinical data based on the parameter ranges, 250,000 samples are created to
serve as input data. The dataset was synthetically generated using normal and uniform distributions
within clinically observed parameter ranges, derived from domain knowledge and clinical literature.
To reflect real-world variability, Gaussian noise was added to selected features. The target variable
Seye was then computed analytically using equation (1), enabling the model to approximate this
function. This setup allows architectural evaluation in a controlled environment prior to testing on
real clinical data;</p>
          <p>3. Calculating the target variable (Seye) using the input data, the target variable Seye is
computed based on a complex nonlinear function (Equation 1). The formula incorporates
logarithmic, exponential, and quadratic dependencies, allowing the realistic modeling of parameter
interactions. The final Seye value serves as a metric for overall eye condition, reflecting the
combined influence of all factors;</p>
          <p>4. Data normalization and partitioning to ensure proper model training, the data was
normalized using MinMaxScaler. The dataset was then split into training (80%) and testing (20%)
subsets. Prior to normalization, outlier detection was performed using multivariate analysis.
Specifically, the Mahalanobis distance was calculated for each observation across the full feature
space (IOP, RQ, BCVA, Tr, VFI, Pperf, Age, and additional factors). Observations with distances
exceeding the 99.5th percentile were classified as multivariate outliers and removed from the
dataset. This ensured the elimination of extreme values that could negatively affect model training,
especially given the nonlinear architecture;</p>
          <p>5. Building the convolutional neural network (CNN) model the data is divided into input and
output subsets, scaled accordingly, and structured into training and test datasets in an 80%/20%
ratio.</p>
          <p>Table 2 presents sample input data, where each row corresponds to an individual set of
measurements. It lists the key parameters used for clinical dataset and subsequent neural network
model training.</p>
          <p>The structure of the convolutional model (Figure 2), developed using an intelligent neural
network, includes:
• input layer with 64 neurons using the ReLU activation function;
• hidden Layer 1 with 128 neurons using ReLU;
• hidden Layer 2 with 64 neurons using ReLU;
• output layer consisting of one neuron with a linear activation function to facilitate
regression-based prediction.</p>
          <p>After data preparation, the neural network training process is initiated on the training dataset.
The key stages include:</p>
          <p>The model consists of an input layer, multiple hidden layers (e.g., three layers with 64, 128, and
64 neurons, respectively) using ReLU activation, and an output layer with linear activation for
regression. This design effectively models nonlinear dependencies between parameters.</p>
          <p>The ReLU (Rectified Linear Unit) activation function was chosen for all hidden layers due to its
simplicity, computational efficiency, and ability to mitigate the vanishing gradient problem, which
is critical for deep neural networks. ReLU introduces non-linearity while preserving gradient flow
for positive inputs, thereby accelerating convergence and improving training performance. For the
output layer, a linear activation function was used, as the task is regression - predicting a
continuous variable (Seye). Linear activation ensures that the network output is not restricted to a
specific range and can represent the full range of possible Seye values, which is necessary for
accurate modeling of clinical variation.</p>
          <p>The Adam optimization algorithm is used for model optimization, while Mean Squared Error
(MSE) is employed as the loss function.</p>
          <p>The model is trained over 100 epochs, achieving high coefficient of determination (R²) values and
minimal errors (RMSE, MAPE). After training, the model is tested on a validation dataset. The
predicted Seye values are compared with actual values calculated using the original formula. A
training progress summary by epochs is presented in Table 3.</p>
          <p>The overall code implements a complete experimental pipeline for predicting eye condition
(S_eye) using deep learning methods. It demonstrates the transition from a theoretical model to its
practical implementation. The code includes:</p>
          <p>1. Definition of input parameters. Clinical value ranges for key ophthalmological parameters
(IOP, RQ, BCVA, etc.) are set;</p>
          <p>2. Formation of clinical data. A dataset of 250,000 samples is created with clinical input
parameter values;</p>
          <p>3. Calculation of the target variable (Seye). The calculate_Seye function incorporates
logarithmic, exponential, and quadratic dependencies, modeling complex nonlinear relationships;
4. Data preprocessing. Data is normalized using MinMaxScaler and split into training (80%)
and testing (20%) subsets;</p>
          <p>5. Model creation and training. A neural network with three hidden layers (64, 128, and 64
neurons, ReLU activation) is trained using Adam optimizer and MSE loss function.</p>
          <p>6. Visualization of the training process. Loss function plots are generated to monitor model
convergence;</p>
          <p>7. Model evaluation. A scatter plot compares predicted and actual S_eye values, and
performance metrics (MAPE, RMSE, R²) are computed;</p>
          <p>8. Model saving and testing. The trained model is serialized and used to predict S_eye on new
input data, with results stored in a CSV file;</p>
          <p>9. Model structure visualization. The plot_model function is used to represent the neural
network architecture.</p>
          <p>The developed code illustrates the entire workflow of the eye condition model - from data
preprocessing to model training, evaluation, and application. It directly reflects the concept
presented in the paper and demonstrates how the theoretical Seye formula can be implemented
using modern deep learning tools to solve the practical task of predicting ophthalmological
conditions.</p>
          <p>The implementation was carried out in Python using Keras with TensorFlow backend. The code
includes all stages of the experimental pipeline: parameter range definition, synthetic data
generation with random sampling and noise, calculation of the target variable Seye using a custom
function, data normalization using MinMaxScaler, and splitting into training and test datasets
(80/20). The neural network was built with three hidden layers using ReLU activation, optimized
with the Adam optimizer, and trained over 100 epochs using Mean Squared Error as the loss
function. Training was performed on a standard laptop (Intel i7, 16 GB RAM) without GPU
acceleration. The average training time was approximately 40 seconds. Visualization modules
(Matplotlib, Seaborn) were used to display learning curves, residual distributions, and scatter plots.
Although the code is not included in the article, its structure directly follows the modeling logic
described and is available upon request.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments</title>
      <p>The error values during neural network training on both the test and training subsets are shown in
Figure 3. This visualization essentially depicts how the error evolves over epochs, indicating that
after the 10th epoch and up to the 100th epoch, no significant error changes occur. The graph
illustrates the relationship between Mean Squared Error (MSE) and training epochs for both the
training (Train) and testing (Test) datasets. Based on the analysis of the graph, the following
conclusions can be drawn:
• the sharp decline in error during the initial epochs indicates that the model quickly adapts
to the data;
• the stabilization of error after the first few epochs suggests that the model reaches a state
where further training does not lead to significant improvements;
• the lack of a significant gap between the training and testing curves indicates that
overfitting is not observed, and the model exhibits good generalization capability.</p>
      <p>The visualization of model training results is further reflected in the distribution of predicted and
actual Seye values (Figure 4).</p>
      <p>The graph is a scatter plot, where the X-axis represents the actual Seye values, and the Y-axis
represents the values predicted by the model. The linear distribution of points along the diagonal
indicates high prediction accuracy. The closer the points are to the y = x line, the more precise the
model's predictions. The absence of significant outliers and dispersion suggests low prediction
error. The high density of points around the diagonal further confirms that the model effectively
approximates the relationship between input data and the target variable Seye. For a sample dataset
IOP = 15, RQ = 2.0, BCVA = 1.5, Tr = 10, VFI = 85, Pperf = 70, age = 45, additional_factor = 0.7:
actual Seye value: 8.0 (calculated from the formula), predicted Seye value: 8.1 (model output). For
another dataset IOP = 30, RQ = 5.5, BCVA = 0.8, Tr = 20, VFI = 60, Pperf = 90, age = 60,
additional_factor = 0.3: actual Seye value: 9.2, predicted Seye value: 9.3.
data, supported by performance metr
Visual analysis (scatter plots) further confirms that the model's accuracy allows it to be effectively
used for comprehensive eye condition assessment.</p>
      <p>The presented graphs (Figures 3 and 4) validate the high accuracy, lack of overfitting, and good
generalization capability of the model. However, for a complete evaluation, an additional metric
such as the coefficient of determination (R²) can be used to quantify how well the model explains
the variance in the target variable. The visualization of these results is presented in Figure 5. The
metric values presented in Figure 5 (MAPE, RMSE, R²) were calculated on the test (validation)
dataset, which was not used during model training. This confirms the mode
and robustness on unseen data.</p>
      <p>The graph shows three performance metrics: MAPE (blue, 0.0020), R² score (green, 0.9988), and
RMSE (red, 0.0178). The low MAPE and RMSE values indicate minimal prediction errors, while the
near-unity R² confirms excellent model fit. These results highlight
ability to capture complex nonlinear patterns. The model was saved, restored, and successfully
validated on new datasets to confirm its generalization capability.A total of 10 prediction
experiments were conducted using the developed model on 10 different input parameter datasets.
The results of Seye predictions on test data are presented in Table 4.</p>
      <p>To analyze residuals, we examine systematic deviations in the model to better understand its
behavior. The histogram (Figure 6) shows that the residuals are symmetrically distributed around
zero without significant skewness, indicating no apparent systematic errors in predicting the Seye
state.</p>
      <p>Additionally, the absence of significant outliers confirms the correct processing of input data and
the stability of the model's predictions.</p>
      <p>The quantile-quantile plot (Figure 7) is used to check the normality of the residual distribution.
The plot shows that most points lie along the diagonal line, indicating that the distribution follows a
normal law.</p>
      <p>Minor deviations at the ends of the graph suggest slight departures from normality, but they are
not critical. Overall, the assumption of residual normality is confirmed, supporting the model's
validity for prediction.</p>
      <p>The residuals vs. predicted values plot (Figure 8) does not reveal any clear patterns, indicating
the absence of systematic model errors. The random distribution of residuals around the zero line
confirms their homoscedasticity (constant variance). This suggests the high quality of the model,
the absence of missing variables, and the lack of significant nonlinearities that were not accounted
for in the neural network architecture.</p>
      <p>The high accuracy of the model (R² = 0.99), low error, and minimal mean absolute percentage
deviation confirm its alignment with actual data. Residual analysis did not reveal significant issues,
indicating the model's reliability in predicting the Seye state.</p>
      <p>Performance is ensured by an optimized architecture using ReLU and Adam. Visualization of the
training process shows a stable error reduction, confirming the absence of significant overfitting.</p>
      <p>Model predictions on test data exhibit high correlation with real values, minimal deviations, and
reliable forecasts.</p>
      <p>Further improvements will focus on enhancing accuracy, optimizing performance, adapting to
real-world data, and expanding functionality. Working with medical data will allow the model to
account for noise and specific correlations, but it will require careful preprocessing.</p>
      <p>Improvements in architecture may involve adding hidden layers, adjusting neuron counts,
applying dropout, and L2 regularization. Additionally, testing gradient boosting methods (XGBoost,
LightGBM) and ensemble modeling is a promising direction.</p>
      <p>Utilizing GPU or TPU will accelerate training, while dimensionality reduction techniques (PCA,
autoencoders) will help reduce computational costs without losing informativeness.</p>
      <p>The experimental section illustrates the entire pipeline, from clinical data and normalization to
neural network training, validation, and evaluation. It presents loss function graphs, scatter plots of
predictions, metric analysis (MAPE, RMSE, R²), and residual distribution analysis.</p>
      <p>The results demonstrate high accuracy, stability, and the absence of overfitting, confirming the
effectiveness of the Seye model for predicting eye conditions.</p>
      <p>Thus, the experimental section outlines the complete workflow, from clinical data creation and
preprocessing to neural network development and training, followed by a detailed comparison of
predicted and actual values. These steps demonstrate that the chosen approach effectively models
complex relationships in ophthalmological data, marking an important step toward developing a
universal tool for eye condition prediction.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Discussion</title>
      <p>This section interprets the obtained results, analyzes model quality, hyperparameter selection, and
compares the proposed approach with existing forecasting methods.</p>
      <p>
        Comparison with traditional forecasting methods. Classical approaches, such as regression
analysis and differential equation-based models, have been widely used for predicting
ophthalmological indicators [
        <xref ref-type="bibr" rid="ref1 ref3 ref5">1, 3, 5</xref>
        ]. However, their main limitation is their inability to accurately
model complex nonlinear relationships between parameters.
      </p>
      <p>
        The proposed neural network model automatically detects hidden dependencies and accounts for
multiple factors affecting eye conditions. Its high acc
its advantages over traditional methods. These findings align with recent studies confirming the
effectiveness of deep learning in ophthalmological disease diagnostics [
        <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
        ].
      </p>
      <p>Considerations on hyperparameter choice. The model used in this study was not subjected to
systematic hyperparameter tuning. Instead, the architecture and parameters (e.g., layer size,
learning rate, activation function) were selected based on common practice and preliminary tests to
ensure stable convergence and low prediction error. A detailed sensitivity analysis remains a
subject for future work. Number of layers and neurons. Increasing the number of neurons enhances
the model's ability to capture complex dependencies but may lead to overfitting. Activation
function. ReLU prevents the vanishing gradient problem and speeds up training. Optimization
parameters. The Adam optimizer demonstrated high efficiency, ensuring fast convergence.</p>
      <p>
        Further hyperparameter tuning using grid search or Bayesian optimization [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] could further
improve model accuracy. Potential model errors and Mitigation strategies. despite the high accuracy
error have been identified:
      </p>
      <p>- edge-case sensitivity. In about 1 2% of test cases, the model exhibits elevated prediction errors
( Seye &gt; 0.1), particularly in extreme parameter ranges (e.g., IOP &gt; 50 mmHg, RQ &gt; 8.0). This
reflects reduced performance in acute or atypical conditions. Enrich the training dataset with rare
pathological cases and apply uncertainty-aware methods (e.g., dropout-based variance estimation);
- limited interaction modeling. Certain parameter combinations (e.g., advanced age with low tear
production and vascular degeneration) lead to mild overestimation of Seye. This indicates
insufficient representation of complex physiological interdependencies. Introduce interaction
layers or attention mechanisms to improve feature coupling;</p>
      <p>- synthetic data limitations. The current model is trained on analytically generated data, which
lacks real-world measurement variability and clinical noise. This restricts its direct applicability in
medical settings. Future work will focus on training and validating the model using real-world
clinical datasets, enabling adaptation to noisy and incomplete inputs.</p>
      <p>It should be emphasized that the current study used a synthetically generated dataset and an
analytically defined target function (Seye), based on physiological modeling. This design allowed
evaluation of neural network architecture under controlled, idealized conditions. Future research
will involve applying the model to real clinical data to assess its diagnostic robustness, expand
interaction modeling, and implement confidence estimation tools. The obtained results confirm that
the proposed neural network model outperforms traditional approaches in both prediction
accuracy and ability to represent nonlinear dependencies. The outlined limitations also
point to specific directions for further optimization and clinical deployment.</p>
    </sec>
    <sec id="sec-6">
      <title>Conclusions and future perspectives</title>
      <p>This study presented a methodological investigation aimed at evaluating a neural network
architecture for approximating a physiologically motivated analytical function (Seye), which models
the overall condition of the eye based on key ophthalmological parameters.</p>
      <p>The main tasks accomplished in this work include:
1. Synthetic data generation and preprocessing. A dataset of 250,000 samples was generated
using clinically observed parameter ranges (IOP, RQ, BCVA, Tr, VFI, Pperf, age, vascular health,
etc.). The Seye values were computed analytically from a predefined function reflecting known
physiological relationships. The dataset was normalized and cleaned for robust model training;
2. Analytical model construction. A semi-empirical formula for Seye was designed using
domain expertise, incorporating nonlinear interactions through logarithmic, exponential, and
polynomial transformations of input parameters. This function served as the basis for supervised
learning;</p>
      <p>3. Neural network design and training. A multi-layer architecture with ReLU activation and a
linear output neuron was trained using the Adam optimizer. The goal was to assess th
ability to approximate the analytical Seye function under idealized, noise-controlled conditions;
4. Validation and performance assessment. The model achieved high R² values and low
RMSE/MAPE when tested against the analytically calculated Seye values, demonstrating strong
approximation capability and confirming the correctness of the selected architecture. Residual
analysis showed no systematic errors or overfitting;</p>
      <p>5. Structural validation for future application. This work does not aim to evaluate real-world
diagnostic accuracy, but rather to establish the readiness of the architecture for subsequent
application to real clinical datasets with inherent noise, variability, and missing values.</p>
      <p>While classical regression and equation-based approaches remain important in ophthalmological
modeling, they are often limited in capturing high-dimensional, nonlinear interactions. The results
confirm that the selected neural network architecture is well-suited for modeling such complexity
even when the target function is semi-empirical in nature. The scientific contribution of this study
lies in providing a validated, computationally efficient architecture that can be used as a foundation
for more advanced clinical applications. This includes integration of real patient data, incorporation
of uncertainty estimation, and development of hybrid models combining machine learning with
physiological modeling. Future Research Directions: application of the architecture to real clinical
datasets to evaluate robustness under data noise and incompleteness; automated hyperparameter
tuning to improve model adaptability; incorporation of explainability and attention mechanisms to
enhance clinical interpretability; development of hybrid frameworks combining deep learning with
classical ophthalmological theory. In conclusion, this study offers a critical validation step in
developing intelligent decision-support tools for ophthalmology. By demonstrating the feasibility of
approximating a physiologically informed model, we lay the groundwork for further clinical
integration and generalization. The presented results reflect structural and algorithmic readiness of
the model, while its clinical utility will be determined through future testing on real-world
ophthalmological data. Generative AI tools, including ChatGPT by OpenAI, were used for example
code generation, phrasing support, and clarity improvement. All scientific concepts, data structures,
and interpretations were developed exclusively by the authors, who take full responsibility for
the integrity of this work.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>The author(s) have not employed any Generative AI tools.
. Liet. Mat.
[23] G. Zhang et al., Multimodal eye imaging, retina characteristics, and psychological assessment
dataset. Sci. Data 11 (2024) 836. doi: 10.1038/s41597-024-03690-6.
[24] L. Kohoutová et al., Toward a unified framework for interpreting machine-learning models in
neuroimaging. Nat. Protoc 15 4 (2020) 1399 1435. doi: 10.1038/s41596-019-0289-5.
[25] H. Kaushik, D. Singh, M. Kaur, H. Alshazly, A. Zaguia, and H. Hamam, Diabetic retinopathy
diagnosis from fundus images using stacked generalization of deep models, IEEE Access, vol. 9,
2021, pp. 108276 108292. doi: 10.1109/ACCESS.2021.3101142.
[26] I. Kansal, V. Khullar, P. Sharma, et al., "Multiple model visual feature embedding and selection
method for an efficient oncular disease classification," Sci. Rep., vol. 15, 2025, pp. 5157. doi:
10.1038/s41598-024-84922-y.
[27] S. Vidivelli, P. Padmakumari, C. Parthiban, et al., "Optimising deep learning models for
ophthalmological disorder classification," Sci. Rep., vol. 15, 2025, p. 3115. doi:
10.1038/s41598024-75867-3.
[28] V. Vychuzhanin et al., "Mathematical model for assessing the functional state of human eye
parameters," Bull. Priazov. State Tech. Univ. Tech. Sci., vol. 49, no. 1, 2024, pp. 6 16. doi:
10.31498/2225-6733.49.1.2024.321178.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Jeong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.-J.</given-names>
            <surname>Hong</surname>
          </string-name>
          , and J.
          <string-name>
            <surname>-H. Han</surname>
          </string-name>
          ,
          <article-title>Review of machine learning applications using retinal fundus images</article-title>
          .
          <source>Diagnostics 12</source>
          <volume>1</volume>
          (
          <year>2022</year>
          )
          <article-title>134</article-title>
          . doi:
          <volume>10</volume>
          .3390/diagnostics12010134.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>K.</given-names>
            <surname>Jin</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Ye</surname>
          </string-name>
          ,
          <article-title>Advances in ophthalmology practice and research</article-title>
          .
          <source>Adv. Ophthalmol. Pract. Res. 2 3</source>
          (
          <issue>2022</issue>
          )
          <article-title>100078</article-title>
          . doi:
          <volume>10</volume>
          .1016/j.aopr.
          <year>2022</year>
          .
          <volume>100078</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bali</surname>
          </string-name>
          and
          <string-name>
            <given-names>V.</given-names>
            <surname>Mansotra</surname>
          </string-name>
          ,
          <article-title>Analysis of deep learning techniques for prediction of eye diseases: A systematic review</article-title>
          .
          <source>Arch. Comput. Methods Eng</source>
          .
          <volume>31</volume>
          (
          <year>2024</year>
          )
          <fpage>487</fpage>
          520. doi:
          <volume>10</volume>
          .1007/s11831-023- 09989-8.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>O.</given-names>
            <surname>Guzun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Zadorozhnyy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Vychuzhanin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Khramenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Velychko</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Korol</surname>
          </string-name>
          ,
          <article-title>A neural network model for predicting the effectiveness of treatment in patients with neovascular glaucoma associated with diabetes mellitus</article-title>
          .
          <source>Rom. J. Ophthalmol. 68 3</source>
          (
          <year>2024</year>
          )
          <fpage>294</fpage>
          300. doi:
          <volume>10</volume>
          .22336/rjo.
          <year>2024</year>
          .
          <volume>53</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>H.</given-names>
            <surname>Bogunovic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Montuoro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Baratsits</surname>
          </string-name>
          , et al.,
          <article-title>Machine learning of the progression of intermediate age-related macular degeneration based on OCT imaging</article-title>
          .
          <source>Invest. Ophthalmol. Vis. Sci. 58</source>
          <volume>6</volume>
          (
          <year>2017</year>
          )
          <fpage>141</fpage>
          150. doi:
          <volume>10</volume>
          .1167/iovs.17-
          <fpage>21789</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Grzybowski</surname>
          </string-name>
          et al.,
          <article-title>Evaluating the efficacy of AI systems in diabetic retinopathy detection: A comparative analysis of Mona DR</article-title>
          and
          <string-name>
            <surname>IDx-DR. Acta Ophthalmol</surname>
          </string-name>
          (
          <year>2024</year>
          ). doi:
          <volume>10</volume>
          .1111/aos.17428.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>E. E.</given-names>
            <surname>Hwang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          , Y. Han,
          <string-name>
            <surname>L</surname>
          </string-name>
          . Jia, and
          <string-name>
            <given-names>J.</given-names>
            <surname>Shan</surname>
          </string-name>
          ,
          <article-title>Utilization of image-based deep learning in multimodal glaucoma detection neural network from a primary patient cohort</article-title>
          .
          <source>Ophthalmol. Sci</source>
          . (
          <year>2025</year>
          ). doi:
          <volume>10</volume>
          .1016/j.xops.
          <year>2025</year>
          .
          <volume>100703</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8] [9]
          <string-name>
            <given-names>and O.</given-names>
            <surname>Zadorozhnyy</surname>
          </string-name>
          ,
          <article-title>Integration of physiological factors into a mathematical model of the human eye condition</article-title>
          .
          <source>Applied Aspects of Information Technology. 87</source>
          .
          <year>2025</year>
          . doi:
          <volume>10</volume>
          .15276/aait.08.
          <year>2025</year>
          .
          <article-title>6</article-title>
          . and
          <string-name>
            <given-names>O.</given-names>
            <surname>Zadorozhnyy</surname>
          </string-name>
          ,
          <article-title>Mathematical modelling of eye condition in glaucoma.: Approaches to parameter analysis and their interactions</article-title>
          .
          <source>Information Technologies and Computer Engineering</source>
          ,
          <volume>22</volume>
          (
          <issue>1</issue>
          ),
          <fpage>9</fpage>
          -
          <lpage>19</lpage>
          .
          <year>2025</year>
          . doi:
          <volume>10</volume>
          .63341/vitce/1.
          <year>2025</year>
          .
          <volume>09</volume>
          .
          <string-name>
            <surname>Rink</surname>
          </string-name>
          .
          <volume>54</volume>
          (
          <year>2013</year>
          ). doi:
          <volume>10</volume>
          .15388/LMR.B.
          <year>2013</year>
          .
          <volume>13</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>R. J.</given-names>
            <surname>Glynn</surname>
          </string-name>
          and
          <string-name>
            <given-names>B.</given-names>
            <surname>Rosner</surname>
          </string-name>
          ,
          <article-title>Regression methods when the eye is the unit of analysis</article-title>
          .
          <source>Ophthalmic Epidemiol. 19 3</source>
          (
          <year>2012</year>
          )
          <fpage>159</fpage>
          165. doi:
          <volume>10</volume>
          .3109/09286586.
          <year>2012</year>
          .
          <volume>674614</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>R.</given-names>
            <surname>Atawi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ayed</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Batran</surname>
          </string-name>
          ,
          <article-title>Traditional eye medicine practice and its determinant factors among ophthalmic patients in the West Bank</article-title>
          .
          <source>J. Public Health Res. 13 2</source>
          (
          <issue>2024</issue>
          )
          <article-title>1 7</article-title>
          . doi:
          <volume>10</volume>
          .1177/22799036241243267.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>D. S. W.</given-names>
            <surname>Ting</surname>
          </string-name>
          et al.,
          <article-title>Artificial intelligence and deep learning in ophthalmology</article-title>
          .
          <source>Br. J. Ophthalmol. 103 2</source>
          (
          <year>2019</year>
          )
          <fpage>167</fpage>
          175. doi:
          <volume>10</volume>
          .1136/bjophthalmol-2018-313173.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>F.</given-names>
            <surname>Du</surname>
          </string-name>
          et al.,
          <article-title>Recognition of eye diseases based on deep neural networks for transfer learning and improved D-S evidence theory</article-title>
          .
          <source>BMC Med</source>
          . Imaging (
          <year>2024</year>
          ).
          <source>doi: 10.1186/s12880-023-01176-2.</source>
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>F. T. J.</given-names>
            <surname>Faria</surname>
          </string-name>
          et al.,
          <article-title>Explainable convolutional neural networks for retinal fundus classification</article-title>
          .
          <source>Electr. Eng. Syst. Sci</source>
          . (
          <year>2024</year>
          ). doi:
          <volume>10</volume>
          .48550/arXiv.2405.07338.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and J.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <article-title>A new type of eye movement model based on recurrent neural networks</article-title>
          .
          <source>Complexity</source>
          (
          <year>2019</year>
          ). doi:
          <volume>10</volume>
          .1155/
          <year>2019</year>
          /8641074.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Rani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Karthikeyini</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C. R.</given-names>
            <surname>Ravi</surname>
          </string-name>
          ,
          <article-title>Eye disease prediction using deep learning and attention on OCT scans</article-title>
          .
          <source>SN Comput. Sci. 5</source>
          (
          <year>2024</year>
          )
          <article-title>1065</article-title>
          . doi:
          <volume>10</volume>
          .1007/s42979-024-03451-7.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>M. H.</given-names>
            <surname>Sarhan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Nasseri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Zapp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Maier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. P.</given-names>
            <surname>Lohmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Navab</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Eslami</surname>
          </string-name>
          ,
          <article-title>Machine learning techniques for ophthalmic data processing: A review</article-title>
          .
          <source>IEEE J. Biomed. Health Inform. 24</source>
          <volume>12</volume>
          (
          <year>2020</year>
          )
          <fpage>3338</fpage>
          3350. doi:
          <volume>10</volume>
          .1109/JBHI.
          <year>2020</year>
          .
          <volume>3012134</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kourosh</surname>
          </string-name>
          et al.,
          <article-title>Integrative therapeutics for ocular surface disorders</article-title>
          .
          <source>Curr. Opin. Allergy Clin. Immunol. 24 5</source>
          (
          <year>2024</year>
          )
          <fpage>397</fpage>
          403. doi:
          <volume>10</volume>
          .1097/ACI.0000000000001024.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>M.</given-names>
            <surname>Moradi</surname>
          </string-name>
          et al.,
          <article-title>Deep ensemble learning for automated non-advanced AMD classification</article-title>
          .
          <source>Comput. Biol. Med</source>
          .
          <volume>154</volume>
          (
          <year>2023</year>
          )
          <article-title>106512</article-title>
          . doi:
          <volume>10</volume>
          .1016/j.compbiomed.
          <year>2022</year>
          .
          <volume>106512</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>E.</given-names>
            <surname>Archana</surname>
          </string-name>
          et al.,
          <article-title>Short analysis of machine learning techniques used for glaucoma detection</article-title>
          ,
          <source>in: Proc. 5th Int. Conf. Smart Syst. Invent. Technol. (ICSSIT)</source>
          ,
          <year>2023</year>
          . doi:
          <volume>10</volume>
          .1109/ICSSIT55814.
          <year>2023</year>
          .
          <volume>10060909</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>L.</given-names>
            <surname>Rueden</surname>
          </string-name>
          et al.,
          <article-title>Combining machine learning and simulation to a hybrid modelling approach</article-title>
          .
          <source>LNCS</source>
          <volume>12080</volume>
          (
          <year>2020</year>
          )
          <fpage>548</fpage>
          560. doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>030</fpage>
          -44584-3_
          <fpage>43</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>